Towards a trustworthy data-driven clinical decision support system: breast cancer use-case
Abstract
Artificial Intelligence (AI) research has emerged as a powerful tool for health-related applications. With the increasing shortage of radiologists and oncologists around the world,
developing an end-to-end AI-based Clinical Decision Support (CDS) system for fatal disease
diagnosis and survivability prediction can have a significant impact on healthcare professionals as well as patients. Such a system uses machine learning algorithms to analyze
medical images and clinical data to detect cancer, estimate its survivability and aid in
treatment planning. We can break the CDS system down into three main components:
the Computer-Aided Diagnosis (CAD), the Computer-Aided Prognosis subsystem (CAP)
and the Computer-Aided Treatment Planning (CATP). The lack of trustworthiness of these
subsystems is still considered a challenge that needs to be addressed in order to increase
their adoption and usefulness in real-world applications. In this thesis, using the breast
cancer use case, we propose new methods and frameworks to address existing challenges
and research gaps in different components of the system to pave the way toward its usage
in clinical practice.
In cancer CAD systems, the first and most important step is to analyze medical images to identify potential tumors in a specific organ. In dense prediction problems like
mass segmentation, preserving the input image resolution plays a crucial role in achieving good performance. However, this resolution is often reduced in current Convolution
Neural Networks (CNN) that are commonly repurposed for this task. In Chapter 3, we
propose a double-dilated convolution module in order to preserve spatial resolution while
having a large receptive field. The proposed module is applied to the tumor segmentation
task in breast cancer mammograms as a proof-of-concept. To address the pixel-level class
imbalance problem in mammogram screenings, different loss functions (i.e., binary crossentropy, weighted cross-entropy, dice loss, and Tversky loss) are evaluated. We address
the lack of transparency in current medical image segmentation models by employing and
quantitatively evaluating different explainability methods (i.e., Grad-CAM, Occlusion Sensitivity, and Activation visualization) for the image segmentation task. Our experimental
analysis shows the effectiveness of the proposed model in increasing the similarity score and
decreasing the miss-detection rate. [...]