4 research outputs found
CELNet: Evidence Localization for Pathology Images using Weakly Supervised Learning
Despite deep convolutional neural networks boost the performance of image
classification and segmentation in digital pathology analysis, they are usually
weak in interpretability for clinical applications or require heavy annotations
to achieve object localization. To overcome this problem, we propose a weakly
supervised learning-based approach that can effectively learn to localize the
discriminative evidence for a diagnostic label from weakly labeled training
data. Experimental results show that our proposed method can reliably pinpoint
the location of cancerous evidence supporting the decision of interest, while
still achieving a competitive performance on glimpse-level and slide-level
histopathologic cancer detection tasks.Comment: Accepted for MICCAI 201
Edge-variational Graph Convolutional Networks for Uncertainty-aware Disease Prediction
There is a rising need for computational models that can complementarily
leverage data of different modalities while investigating associations between
subjects for population-based disease analysis. Despite the success of
convolutional neural networks in representation learning for imaging data, it
is still a very challenging task. In this paper, we propose a generalizable
framework that can automatically integrate imaging data with non-imaging data
in populations for uncertainty-aware disease prediction. At its core is a
learnable adaptive population graph with variational edges, which we
mathematically prove that it is optimizable in conjunction with graph
convolutional neural networks. To estimate the predictive uncertainty related
to the graph topology, we propose the novel concept of Monte-Carlo edge
dropout. Experimental results on four databases show that our method can
consistently and significantly improve the diagnostic accuracy for Autism
spectrum disorder, Alzheimer's disease, and ocular diseases, indicating its
generalizability in leveraging multimodal data for computer-aided diagnosis.Comment: Accepted to MICCAI 202
Gigapixel Histopathological Image Analysis using Attention-based Neural Networks
Although CNNs are widely considered as the state-of-the-art models in various
applications of image analysis, one of the main challenges still open is the
training of a CNN on high resolution images. Different strategies have been
proposed involving either a rescaling of the image or an individual processing
of parts of the image. Such strategies cannot be applied to images, such as
gigapixel histopathological images, for which a high reduction in resolution
inherently effects a loss of discriminative information, and in respect of
which the analysis of single parts of the image suffers from a lack of global
information or implies a high workload in terms of annotating the training
images in such a way as to select significant parts. We propose a method for
the analysis of gigapixel histopathological images solely by using weak
image-level labels. In particular, two analysis tasks are taken into account: a
binary classification and a prediction of the tumor proliferation score. Our
method is based on a CNN structure consisting of a compressing path and a
learning path. In the compressing path, the gigapixel image is packed into a
grid-based feature map by using a residual network devoted to the feature
extraction of each patch into which the image has been divided. In the learning
path, attention modules are applied to the grid-based feature map, taking into
account spatial correlations of neighboring patch features to find regions of
interest, which are then used for the final whole slide analysis. Our method
integrates both global and local information, is flexible with regard to the
size of the input images and only requires weak image-level labels. Comparisons
with different methods of the state-of-the-art on two well known datasets,
Camelyon16 and TUPAC16, have been made to confirm the validity of the proposed
model.Comment: The manuscript was submitted to a peer-review journal on January 27t
Explainable artificial intelligence (XAI) in deep learning-based medical image analysis
With an increase in deep learning-based methods, the call for explainability
of such methods grows, especially in high-stakes decision making areas such as
medical image analysis. This survey presents an overview of eXplainable
Artificial Intelligence (XAI) used in deep learning-based medical image
analysis. A framework of XAI criteria is introduced to classify deep
learning-based medical image analysis methods. Papers on XAI techniques in
medical image analysis are then surveyed and categorized according to the
framework and according to anatomical location. The paper concludes with an
outlook of future opportunities for XAI in medical image analysis.Comment: Submitted for publication. Comments welcome by email to first autho