45 research outputs found
Factorised spatial representation learning: application in semi-supervised myocardial segmentation
The success and generalisation of deep learning algorithms heavily depend on
learning good feature representations. In medical imaging this entails
representing anatomical information, as well as properties related to the
specific imaging setting. Anatomical information is required to perform further
analysis, whereas imaging information is key to disentangle scanner variability
and potential artefacts. The ability to factorise these would allow for
training algorithms only on the relevant information according to the task. To
date, such factorisation has not been attempted. In this paper, we propose a
methodology of latent space factorisation relying on the cycle-consistency
principle. As an example application, we consider cardiac MR segmentation,
where we separate information related to the myocardium from other features
related to imaging and surrounding substructures. We demonstrate the proposed
method's utility in a semi-supervised setting: we use very few labelled images
together with many unlabelled images to train a myocardium segmentation neural
network. Specifically, we achieve comparable performance to fully supervised
networks using a fraction of labelled images in experiments on ACDC and a
dataset from Edinburgh Imaging Facility QMRI. Code will be made available at
https://github.com/agis85/spatial_factorisation.Comment: Accepted in MICCAI 201
Learning to synthesise the ageing brain without longitudinal data
How will my face look when I get older? Or, for a more challenging question:
How will my brain look when I get older? To answer this question one must
devise (and learn from data) a multivariate auto-regressive function which
given an image and a desired target age generates an output image. While
collecting data for faces may be easier, collecting longitudinal brain data is
not trivial. We propose a deep learning-based method that learns to simulate
subject-specific brain ageing trajectories without relying on longitudinal
data. Our method synthesises images conditioned on two factors: age (a
continuous variable), and status of Alzheimer's Disease (AD, an ordinal
variable). With an adversarial formulation we learn the joint distribution of
brain appearance, age and AD status, and define reconstruction losses to
address the challenging problem of preserving subject identity. We compare with
several benchmarks using two widely used datasets. We evaluate the quality and
realism of synthesised images using ground-truth longitudinal data and a
pre-trained age predictor. We show that, despite the use of cross-sectional
data, our model learns patterns of gray matter atrophy in the middle temporal
gyrus in patients with AD. To demonstrate generalisation ability, we train on
one dataset and evaluate predictions on the other. In conclusion, our model
shows an ability to separate age, disease influence and anatomy using only 2D
cross-sectional data that should be useful in large studies into
neurodegenerative disease, that aim to combine several data sources. To
facilitate such future studies by the community at large our code is made
available at https://github.com/xiat0616/BrainAgeing
Contrastive learning for view classification of echocardiograms
Analysis of cardiac ultrasound images is commonly performed in routine clinical practice for quantification of cardiac function. Its increasing automation frequently employs deep learning networks that are trained to predict disease or detect image features. However, such models are extremely data-hungry and training requires labelling of many thousands of images by experienced clinicians. Here we propose the use of contrastive learning to mitigate the labelling bottleneck. We train view classification models for imbalanced cardiac ultrasound datasets and show improved performance for views/classes for which minimal labelled data is available. Compared to a naïve baseline model, we achieve an improvement in F1 score of up to 26% in those views while maintaining state-of-the-art performance for the views with sufficiently many labelled training observations
INSIDE: Steering Spatial Attention with Non-Imaging Information in CNNs
We consider the problem of integrating non-imaging information into
segmentation networks to improve performance. Conditioning layers such as FiLM
provide the means to selectively amplify or suppress the contribution of
different feature maps in a linear fashion. However, spatial dependency is
difficult to learn within a convolutional paradigm. In this paper, we propose a
mechanism to allow for spatial localisation conditioned on non-imaging
information, using a feature-wise attention mechanism comprising a
differentiable parametrised function (e.g. Gaussian), prior to applying the
feature-wise modulation. We name our method INstance modulation with SpatIal
DEpendency (INSIDE). The conditioning information might comprise any factors
that relate to spatial or spatio-temporal information such as lesion location,
size, and cardiac cycle phase. Our method can be trained end-to-end and does
not require additional supervision. We evaluate the method on two datasets: a
new CLEVR-Seg dataset where we segment objects based on location, and the ACDC
dataset conditioned on cardiac phase and slice location within the volume. Code
and the CLEVR-Seg dataset are available at https://github.com/jacenkow/inside.Comment: Accepted at International Conference on Medical Image Computing and
Computer Assisted Intervention (MICCAI) 202
Disentangled Representations for Domain-generalized Cardiac Segmentation
Robust cardiac image segmentation is still an open challenge due to the
inability of the existing methods to achieve satisfactory performance on unseen
data of different domains. Since the acquisition and annotation of medical data
are costly and time-consuming, recent work focuses on domain adaptation and
generalization to bridge the gap between data from different populations and
scanners. In this paper, we propose two data augmentation methods that focus on
improving the domain adaptation and generalization abilities of
state-to-the-art cardiac segmentation models. In particular, our "Resolution
Augmentation" method generates more diverse data by rescaling images to
different resolutions within a range spanning different scanner protocols.
Subsequently, our "Factor-based Augmentation" method generates more diverse
data by projecting the original samples onto disentangled latent spaces, and
combining the learned anatomy and modality factors from different domains. Our
extensive experiments demonstrate the importance of efficient adaptation
between seen and unseen domains, as well as model generalization ability, to
robust cardiac image segmentation.Comment: Accepted by STACOM 202