48 research outputs found
Generative-Discriminative Low Rank Decomposition for Medical Imaging Applications
In this thesis, we propose a method that can be used to extract biomarkers from medical images toward early diagnosis of abnormalities. Surge of demand for biomarkers and availability of medical images in the recent years call for accurate, repeatable, and interpretable approaches for extracting meaningful imaging features. However, extracting such information from medical images is a challenging task because the number of pixels (voxels) in a typical image is in order of millions while even a large sample-size in medical image dataset does not usually exceed a few hundred. Nevertheless, depending on the nature of an abnormality, only a parsimonious subset of voxels is typically relevant to the disease; therefore various notions of sparsity are exploited in this thesis to improve the generalization performance of the prediction task.
We propose a novel discriminative dimensionality reduction method that yields good classification performance on various datasets without compromising the clinical interpretability of the results. This is achieved by combining the modelling strength of generative learning framework and the classification performance of discriminative learning paradigm. Clinical interpretability can be viewed as an additional measure of evaluation and is also helpful in designing methods that account for the clinical prior such as association of certain areas in a brain to a particular cognitive task or connectivity of some brain regions via neural fibres.
We formulate our method as a large-scale optimization problem to solve a constrained matrix factorization. Finding an optimal solution of the large-scale matrix factorization renders off-the-shelf solver computationally prohibitive; therefore, we designed an efficient algorithm based on the proximal method to address the computational bottle-neck of the optimization problem. Our formulation is readily extended for different scenarios such as cases where a large cohort of subjects has uncertain or no class labels (semi-supervised learning) or a case where each subject has a battery of imaging channels (multi-channel), \etc. We show that by using various notions of sparsity as feasible sets of the optimization problem, we can encode different forms of prior knowledge ranging from brain parcellation to brain connectivity
Distilling BlackBox to Interpretable models for Efficient Transfer Learning
Building generalizable AI models is one of the primary challenges in the
healthcare domain. While radiologists rely on generalizable descriptive rules
of abnormality, Neural Network (NN) models suffer even with a slight shift in
input distribution (\eg scanner type). Fine-tuning a model to transfer
knowledge from one domain to another requires a significant amount of labeled
data in the target domain. In this paper, we develop an interpretable model
that can be efficiently fine-tuned to an unseen target domain with minimal
computational cost. We assume the interpretable component of NN to be
approximately domain-invariant. However, interpretable models typically
underperform compared to their Blackbox (BB) variants. We start with a BB in
the source domain and distill it into a \emph{mixture} of shallow interpretable
models using human-understandable concepts. As each interpretable model covers
a subset of data, a mixture of interpretable models achieves comparable
performance as BB. Further, we use the pseudo-labeling technique from
semi-supervised learning (SSL) to learn the concept classifier in the target
domain, followed by fine-tuning the interpretable models in the target domain.
We evaluate our model using a real-life large-scale chest-X-ray (CXR)
classification dataset. The code is available at:
\url{https://github.com/batmanlab/MICCAI-2023-Route-interpret-repeat-CXRs}.Comment: MICCAI, 2023, Early accep
Context Matters: Graph-based Self-supervised Representation Learning for Medical Images
Supervised learning method requires a large volume of annotated datasets.
Collecting such datasets is time-consuming and expensive. Until now, very few
annotated COVID-19 imaging datasets are available. Although self-supervised
learning enables us to bootstrap the training by exploiting unlabeled data, the
generic self-supervised methods for natural images do not sufficiently
incorporate the context. For medical images, a desirable method should be
sensitive enough to detect deviation from normal-appearing tissue of each
anatomical region; here, anatomy is the context. We introduce a novel approach
with two levels of self-supervised representation learning objectives: one on
the regional anatomical level and another on the patient-level. We use graph
neural networks to incorporate the relationship between different anatomical
regions. The structure of the graph is informed by anatomical correspondences
between each patient and an anatomical atlas. In addition, the graph
representation has the advantage of handling any arbitrarily sized image in
full resolution. Experiments on large-scale Computer Tomography (CT) datasets
of lung images show that our approach compares favorably to baseline methods
that do not account for the context. We use the learnt embedding to quantify
the clinical progression of COVID-19 and show that our method generalizes well
to COVID-19 patients from different hospitals. Qualitative results suggest that
our model can identify clinically relevant regions in the images.Comment: Accepted to AAAI 202