75 research outputs found
Joint Segmentation and Uncertainty Visualization of Retinal Layers in Optical Coherence Tomography Images using Bayesian Deep Learning
Optical coherence tomography (OCT) is commonly used to analyze retinal layers
for assessment of ocular diseases. In this paper, we propose a method for
retinal layer segmentation and quantification of uncertainty based on Bayesian
deep learning. Our method not only performs end-to-end segmentation of retinal
layers, but also gives the pixel wise uncertainty measure of the segmentation
output. The generated uncertainty map can be used to identify erroneously
segmented image regions which is useful in downstream analysis. We have
validated our method on a dataset of 1487 images obtained from 15 subjects (OCT
volumes) and compared it against the state-of-the-art segmentation algorithms
that does not take uncertainty into account. The proposed uncertainty based
segmentation method results in comparable or improved performance, and most
importantly is more robust against noise
Shallow vs deep learning architectures for white matter lesion segmentation in the early stages of multiple sclerosis
In this work, we present a comparison of a shallow and a deep learning
architecture for the automated segmentation of white matter lesions in MR
images of multiple sclerosis patients. In particular, we train and test both
methods on early stage disease patients, to verify their performance in
challenging conditions, more similar to a clinical setting than what is
typically provided in multiple sclerosis segmentation challenges. Furthermore,
we evaluate a prototype naive combination of the two methods, which refines the
final segmentation. All methods were trained on 32 patients, and the evaluation
was performed on a pure test set of 73 cases. Results show low lesion-wise
false positives (30%) for the deep learning architecture, whereas the shallow
architecture yields the best Dice coefficient (63%) and volume difference
(19%). Combining both shallow and deep architectures further improves the
lesion-wise metrics (69% and 26% lesion-wise true and false positive rate,
respectively).Comment: Accepted to the MICCAI 2018 Brain Lesion (BrainLes) worksho
Applying an Open-Source Segmentation Algorithm to Different OCT Devices in Multiple Sclerosis Patients and Healthy Controls: Implications for Clinical Trials
Background. The lack of segmentation algorithms operative across optical coherence tomography (OCT) platforms hinders utility of retinal layer measures in MS trials. Objective. To determine cross-sectional and longitudinal agreement of retinal layer thicknesses derived from an open-source, fully-automated, segmentation algorithm, applied to two spectral-domain OCT devices. Methods. Cirrus HD-OCT and Spectralis OCT macular scans from 68 MS patients and 22 healthy controls were segmented. A longitudinal cohort comprising 51 subjects (mean follow-up: 1.4 ± 0.9 years) was also examined. Bland-Altman analyses and interscanner agreement indices were utilized to assess agreement between scanners. Results. Low mean differences (−2.16 to 0.26 μm) and narrow limits of agreement (LOA) were noted for ganglion cell and inner and outer nuclear layer thicknesses cross-sectionally. Longitudinally we found low mean differences (−0.195 to 0.21 μm) for changes in all layers, with wider LOA. Comparisons of rate of change in layer thicknesses over time revealed consistent results between the platforms. Conclusions. Retinal thickness measures for the majority of the retinal layers agree well cross-sectionally and longitudinally between the two scanners at the cohort level, with greater variability at the individual level. This open-source segmentation algorithm enables combining data from different OCT platforms, broadening utilization of OCT as an outcome measure in MS trials
Intensity Inhomogeneity Correction of SD-OCT Data Using Macular Flatspace
Images of the retina acquired using optical coherence tomography (OCT) often suffer from intensity inhomogeneity problems that degrade both the quality of the images and the performance of automated algorithms utilized to measure structural changes. This intensity variation has many causes, including off-axis acquisition, signal attenuation, multi-frame averaging, and vignetting, making it difficult to correct the data in a fundamental way. This paper presents a method for inhomogeneity correction by acting to reduce the variability of intensities within each layer. In particular, the N3 algorithm, which is popular in neuroimage analysis, is adapted to work for OCT data. N3 works by sharpening the intensity histogram, which reduces the variation of intensities within different classes. To apply it here, the data are first converted to a standardized space called macular flat space (MFS). MFS allows the intensities within each layer to be more easily normalized by removing the natural curvature of the retina. N3 is then run on the MFS data using a modified smoothing model, which improves the efficiency of the original algorithm. We show that our method more accurately corrects gain fields on synthetic OCT data when compared to running N3 on non-flattened data. It also reduces the overall variability of the intensities within each layer, without sacrificing contrast between layers, and improves the performance of registration between OCT images
Improved inter-scanner MS lesion segmentation by adversarial training on longitudinal data
The evaluation of white matter lesion progression is an important biomarker
in the follow-up of MS patients and plays a crucial role when deciding the
course of treatment. Current automated lesion segmentation algorithms are
susceptible to variability in image characteristics related to MRI scanner or
protocol differences. We propose a model that improves the consistency of MS
lesion segmentations in inter-scanner studies. First, we train a CNN base model
to approximate the performance of icobrain, an FDA-approved clinically
available lesion segmentation software. A discriminator model is then trained
to predict if two lesion segmentations are based on scans acquired using the
same scanner type or not, achieving a 78% accuracy in this task. Finally, the
base model and the discriminator are trained adversarially on multi-scanner
longitudinal data to improve the inter-scanner consistency of the base model.
The performance of the models is evaluated on an unseen dataset containing
manual delineations. The inter-scanner variability is evaluated on test-retest
data, where the adversarial network produces improved results over the base
model and the FDA-approved solution.Comment: MICCAI BrainLes 2019 Worksho
Test-time Unsupervised Domain Adaptation
Convolutional neural networks trained on publicly available medical imaging
datasets (source domain) rarely generalise to different scanners or acquisition
protocols (target domain). This motivates the active field of domain
adaptation. While some approaches to the problem require labeled data from the
target domain, others adopt an unsupervised approach to domain adaptation
(UDA). Evaluating UDA methods consists of measuring the model's ability to
generalise to unseen data in the target domain. In this work, we argue that
this is not as useful as adapting to the test set directly. We therefore
propose an evaluation framework where we perform test-time UDA on each subject
separately. We show that models adapted to a specific target subject from the
target domain outperform a domain adaptation method which has seen more data of
the target domain but not this specific target subject. This result supports
the thesis that unsupervised domain adaptation should be used at test-time,
even if only using a single target-domain subjectComment: Accepted at MICCAI 202
Rapid Brain Meninges Surface Reconstruction with Layer Topology Guarantee
The meninges, located between the skull and brain, are composed of three
membrane layers: the pia, the arachnoid, and the dura. Reconstruction of these
layers can aid in studying volume differences between patients with
neurodegenerative diseases and normal aging subjects. In this work, we use
convolutional neural networks (CNNs) to reconstruct surfaces representing
meningeal layer boundaries from magnetic resonance (MR) images. We first use
the CNNs to predict the signed distance functions (SDFs) representing these
surfaces while preserving their anatomical ordering. The marching cubes
algorithm is then used to generate continuous surface representations; both the
subarachnoid space (SAS) and the intracranial volume (ICV) are computed from
these surfaces. The proposed method is compared to a state-of-the-art
deformable model-based reconstruction method, and we show that our method can
reconstruct smoother and more accurate surfaces using less computation time.
Finally, we conduct experiments with volumetric analysis on both subjects with
multiple sclerosis and healthy controls. For healthy and MS subjects, ICVs and
SAS volumes are found to be significantly correlated to sex (p<0.01) and age
(p<0.03) changes, respectively.Comment: ISBI 2023 Ora
Bio-inspired Attentive Segmentation of Retinal OCT Imaging
Albeit optical coherence imaging (OCT) is widely used to assess ophthalmic pathologies, localization of intra-retinal boundaries suffers from erroneous segmentations due to image artifacts or topological abnormalities. Although deep learning-based methods have been effectively applied in OCT imaging, accurate automated layer segmentation remains a challenging task, with the flexibility and precision of most methods being highly constrained. In this paper, we propose a novel method to segment all retinal layers, tailored to the bio-topological OCT geometry. In addition to traditional learning of shift-invariant features, our method learns in selected pixels horizontally and vertically, exploiting the orientation of the extracted features. In this way, the most discriminative retinal features are generated in a robust manner, while long-range pixel dependencies across spatial locations are efficiently captured. To validate the effectiveness and generalisation of our method, we implement three sets of networks based on different backbone models. Results on three independent studies show that our methodology consistently produces more accurate segmentations than state-of-the-art networks, and shows better precision and agreement with ground truth. Thus, our method not only improves segmentation, but also enhances the statistical power of clinical trials with layer thickness change outcomes
HACA3: A Unified Approach for Multi-site MR Image Harmonization
The lack of standardization is a prominent issue in magnetic resonance (MR)
imaging. This often causes undesired contrast variations due to differences in
hardware and acquisition parameters. In recent years, MR harmonization using
image synthesis with disentanglement has been proposed to compensate for the
undesired contrast variations. Despite the success of existing methods, we
argue that three major improvements can be made. First, most existing methods
are built upon the assumption that multi-contrast MR images of the same subject
share the same anatomy. This assumption is questionable since different MR
contrasts are specialized to highlight different anatomical features. Second,
these methods often require a fixed set of MR contrasts for training (e.g.,
both Tw-weighted and T2-weighted images must be available), which limits their
applicability. Third, existing methods generally are sensitive to imaging
artifacts. In this paper, we present a novel approach, Harmonization with
Attention-based Contrast, Anatomy, and Artifact Awareness (HACA3), to address
these three issues. We first propose an anatomy fusion module that enables
HACA3 to respect the anatomical differences between MR contrasts. HACA3 is also
robust to imaging artifacts and can be trained and applied to any set of MR
contrasts. Experiments show that HACA3 achieves state-of-the-art performance
under multiple image quality metrics. We also demonstrate the applicability of
HACA3 on downstream tasks with diverse MR datasets acquired from 21 sites with
different field strengths, scanner platforms, and acquisition protocols
A Longitudinal Method for Simultaneous Whole-Brain and Lesion Segmentation in Multiple Sclerosis
In this paper we propose a novel method for the segmentation of longitudinal
brain MRI scans of patients suffering from Multiple Sclerosis. The method
builds upon an existing cross-sectional method for simultaneous whole-brain and
lesion segmentation, introducing subject-specific latent variables to encourage
temporal consistency between longitudinal scans. It is very generally
applicable, as it does not make any prior assumptions on the scanner, the MRI
protocol, or the number and timing of longitudinal follow-up scans. Preliminary
experiments on three longitudinal datasets indicate that the proposed method
produces more reliable segmentations and detects disease effects better than
the cross-sectional method it is based upon
- …