6 research outputs found
A Survey on Deep Learning in Medical Image Registration: New Technologies, Uncertainty, Evaluation Metrics, and Beyond
Over the past decade, deep learning technologies have greatly advanced the
field of medical image registration. The initial developments, such as
ResNet-based and U-Net-based networks, laid the groundwork for deep
learning-driven image registration. Subsequent progress has been made in
various aspects of deep learning-based registration, including similarity
measures, deformation regularizations, and uncertainty estimation. These
advancements have not only enriched the field of deformable image registration
but have also facilitated its application in a wide range of tasks, including
atlas construction, multi-atlas segmentation, motion estimation, and 2D-3D
registration. In this paper, we present a comprehensive overview of the most
recent advancements in deep learning-based image registration. We begin with a
concise introduction to the core concepts of deep learning-based image
registration. Then, we delve into innovative network architectures, loss
functions specific to registration, and methods for estimating registration
uncertainty. Additionally, this paper explores appropriate evaluation metrics
for assessing the performance of deep learning models in registration tasks.
Finally, we highlight the practical applications of these novel techniques in
medical imaging and discuss the future prospects of deep learning-based image
registration
Fetal-BET: Brain Extraction Tool for Fetal MRI
Fetal brain extraction is a necessary first step in most computational fetal
brain MRI pipelines. However, it has been a very challenging task due to
non-standard fetal head pose, fetal movements during examination, and vastly
heterogeneous appearance of the developing fetal brain and the neighboring
fetal and maternal anatomy across various sequences and scanning conditions.
Development of a machine learning method to effectively address this task
requires a large and rich labeled dataset that has not been previously
available. As a result, there is currently no method for accurate fetal brain
extraction on various fetal MRI sequences. In this work, we first built a large
annotated dataset of approximately 72,000 2D fetal brain MRI images. Our
dataset covers the three common MRI sequences including T2-weighted,
diffusion-weighted, and functional MRI acquired with different scanners.
Moreover, it includes normal and pathological brains. Using this dataset, we
developed and validated deep learning methods, by exploiting the power of the
U-Net style architectures, the attention mechanism, multi-contrast feature
learning, and data augmentation for fast, accurate, and generalizable automatic
fetal brain extraction. Our approach leverages the rich information from
multi-contrast (multi-sequence) fetal MRI data, enabling precise delineation of
the fetal brain structures. Evaluations on independent test data show that our
method achieves accurate brain extraction on heterogeneous test data acquired
with different scanners, on pathological brains, and at various gestational
stages. This robustness underscores the potential utility of our deep learning
model for fetal brain imaging and image analysis.Comment: 10 pages, 6 figures, 2 TABLES, This work has been submitted to the
IEEE Transactions on Medical Imaging for possible publication. Copyright may
be transferred without notice, after which this version may no longer be
accessibl
A Multi-scale Learning of Data-driven and Anatomically Constrained Image Registration for Adult and Fetal Echo Images
Temporal echo image registration is a basis for clinical quantifications such
as cardiac motion estimation, myocardial strain assessments, and stroke volume
quantifications. Deep learning image registration (DLIR) is consistently
accurate, requires less computing effort, and has shown encouraging results in
earlier applications. However, we propose that a greater focus on the warped
moving image's anatomic plausibility and image quality can support robust DLIR
performance. Further, past implementations have focused on adult echo, and
there is an absence of DLIR implementations for fetal echo. We propose a
framework combining three strategies for DLIR for both fetal and adult echo:
(1) an anatomic shape-encoded loss to preserve physiological myocardial and
left ventricular anatomical topologies in warped images; (2) a data-driven loss
that is trained adversarially to preserve good image texture features in warped
images; and (3) a multi-scale training scheme of a data-driven and anatomically
constrained algorithm to improve accuracy. Our experiments show that the
shape-encoded loss and the data-driven adversarial loss are strongly correlated
to good anatomical topology and image textures, respectively. They improve
different aspects of registration performance in a non-overlapping way,
justifying their combination. We show that these strategies can provide
excellent registration results in both adult and fetal echo using the publicly
available CAMUS adult echo dataset and our private multi-demographic fetal echo
dataset, despite fundamental distinctions between adult and fetal echo images.
Our approach also outperforms traditional non-DL gold standard registration
approaches, including Optical Flow and Elastix. Registration improvements could
also be translated to more accurate and precise clinical quantification of
cardiac ejection fraction, demonstrating a potential for translation
How good is good enough? Strategies for dealing with unreliable segmentation annotations of medical data
Medical image segmentation is an essential topic in computer vision and medical image analysis, because it enables the precise and accurate segmentation of organs and lesions for healthcare applications. Deep learning has dominated in medical image segmentation due to increasingly powerful computational resources, successful neural network architecture engineering, and access to large amounts of medical imaging data with high-quality annotations. However, annotating medical imaging data is time-consuming and expensive, and sometimes the annotations are unreliable.
This DPhil thesis presents a comprehensive study that explores deep learning techniques in medical image segmentation under various challenging situations of unreliable medical imaging data. These situations include: (1) conventional supervised learning to tackle comprehensive data annotation with full dense masks, (2) semi-supervised learning to tackle partial data annotation with full dense masks, (3) noise-robust learning to tackle comprehensive data annotation with noisy dense masks, and (4) weakly-supervised learning to tackle comprehensive data annotation with sketchy contours for network training.
The proposed medical image segmentation strategies improve deep learning techniques to effectively address a series of challenges in medical image analysis, including limited annotated data, noisy annotations, and sparse annotations. These advancements aim to bring deep learning techniques of medical image analysis into practical clinical scenarios. By overcoming these challenges, the strategies establish a more robust and reliable application of deep learning methods which is valuable for improving diagnostic precision and patient care outcomes in real-world clinical environments
Fluorescence microscopy image analysis of retinal neurons using deep learning
An essential goal of neuroscience is to understand the brain by simultaneously identifying, measuring, and analyzing activity from individual cells within a neural population in live brain tissue. Analyzing fluorescence microscopy (FM) images in real-time with computational algorithms is essential for achieving this goal. Deep learning techniques have shown promise in this area, but face domain-specific challenges due to limited training data, significant amounts of voxel noise in FM images, and thin structures present in large 3D images. In this thesis, I address these issues by introducing a novel deep learning pipeline to analyze static FM images of neurons with minimal data requirements and demonstrate the pipeline’s ability to segment neurons from low signal-to-noise ratio FM images with few training samples. The first step of this pipeline employs a Generative Adversarial Network (GAN) equipped to learn imaging properties from a small set of static FM images acquired for a given neuroscientific experiment. Operating like an actual microscope, our fully-trained GAN can then generate realistic static FM images from volumetric reconstructions of neurons with added control over the intensity and noise of the generated images. For the second step in our pipeline, a novel segmentation network is trained on GAN-generated images with reconstructed neurons serving as “gold standard” ground truths. While training on a large dataset of FM images is optimal, a 15\% improvement in neuron segmentation accuracy from noisy FM images is shown when architectures are fine-tuned only on a small subsample of real image data. To evaluate the overall feasibility of our pipeline and the utility of generated images, 2 novel synthetic and 3 newly acquired FM image datasets are introduced along with a new evaluation protocol for FM image ”realness” that incorporates content, texture, and expert opinion metrics. While this pipeline's primary application is to segment neurons from highly noisy FM images, its utility can be extended to automate other FM tasks such as synapse identification, neuron classification, or super-resolution