256 research outputs found
AFFIRM: Affinity Fusion-based Framework for Iteratively Random Motion correction of multi-slice fetal brain MRI
Multi-slice magnetic resonance images of the fetal brain are usually
contaminated by severe and arbitrary fetal and maternal motion. Hence, stable
and robust motion correction is necessary to reconstruct high-resolution 3D
fetal brain volume for clinical diagnosis and quantitative analysis. However,
the conventional registration-based correction has a limited capture range and
is insufficient for detecting relatively large motions. Here, we present a
novel Affinity Fusion-based Framework for Iteratively Random Motion (AFFIRM)
correction of the multi-slice fetal brain MRI. It learns the sequential motion
from multiple stacks of slices and integrates the features between 2D slices
and reconstructed 3D volume using affinity fusion, which resembles the
iterations between slice-to-volume registration and volumetric reconstruction
in the regular pipeline. The method accurately estimates the motion regardless
of brain orientations and outperforms other state-of-the-art learning-based
methods on the simulated motion-corrupted data, with a 48.4% reduction of mean
absolute error for rotation and 61.3% for displacement. We then incorporated
AFFIRM into the multi-resolution slice-to-volume registration and tested it on
the real-world fetal MRI scans at different gestation stages. The results
indicated that adding AFFIRM to the conventional pipeline improved the success
rate of fetal brain super-resolution reconstruction from 77.2% to 91.9%
A Survey on Deep Learning in Medical Image Analysis
Deep learning algorithms, in particular convolutional networks, have rapidly
become a methodology of choice for analyzing medical images. This paper reviews
the major deep learning concepts pertinent to medical image analysis and
summarizes over 300 contributions to the field, most of which appeared in the
last year. We survey the use of deep learning for image classification, object
detection, segmentation, registration, and other tasks and provide concise
overviews of studies per application area. Open challenges and directions for
future research are discussed.Comment: Revised survey includes expanded discussion section and reworked
introductory section on common deep architectures. Added missed papers from
before Feb 1st 201
Fully Convolutional Slice-to-Volume Reconstruction for Single-Stack MRI
In magnetic resonance imaging (MRI), slice-to-volume reconstruction (SVR)
refers to computational reconstruction of an unknown 3D magnetic resonance
volume from stacks of 2D slices corrupted by motion. While promising, current
SVR methods require multiple slice stacks for accurate 3D reconstruction,
leading to long scans and limiting their use in time-sensitive applications
such as fetal fMRI. Here, we propose a SVR method that overcomes the
shortcomings of previous work and produces state-of-the-art reconstructions in
the presence of extreme inter-slice motion. Inspired by the recent success of
single-view depth estimation methods, we formulate SVR as a single-stack motion
estimation task and train a fully convolutional network to predict a motion
stack for a given slice stack, producing a 3D reconstruction as a byproduct of
the predicted motion. Extensive experiments on the SVR of adult and fetal
brains demonstrate that our fully convolutional method is twice as accurate as
previous SVR methods. Our code is available at github.com/seannz/svr.Comment: Accepted to CVPR 202
Deep Learning in Cardiology
The medical field is creating large amount of data that physicians are unable
to decipher and use efficiently. Moreover, rule-based expert systems are
inefficient in solving complicated medical tasks or for creating insights using
big data. Deep learning has emerged as a more accurate and effective technology
in a wide range of medical problems such as diagnosis, prediction and
intervention. Deep learning is a representation learning method that consists
of layers that transform the data non-linearly, thus, revealing hierarchical
relationships and structures. In this review we survey deep learning
application papers that use structured data, signal and imaging modalities from
cardiology. We discuss the advantages and limitations of applying deep learning
in cardiology that also apply in medicine in general, while proposing certain
directions as the most viable for clinical use.Comment: 27 pages, 2 figures, 10 table
3D Deep Learning on Medical Images: A Review
The rapid advancements in machine learning, graphics processing technologies
and availability of medical imaging data has led to a rapid increase in use of
deep learning models in the medical domain. This was exacerbated by the rapid
advancements in convolutional neural network (CNN) based architectures, which
were adopted by the medical imaging community to assist clinicians in disease
diagnosis. Since the grand success of AlexNet in 2012, CNNs have been
increasingly used in medical image analysis to improve the efficiency of human
clinicians. In recent years, three-dimensional (3D) CNNs have been employed for
analysis of medical images. In this paper, we trace the history of how the 3D
CNN was developed from its machine learning roots, give a brief mathematical
description of 3D CNN and the preprocessing steps required for medical images
before feeding them to 3D CNNs. We review the significant research in the field
of 3D medical imaging analysis using 3D CNNs (and its variants) in different
medical areas such as classification, segmentation, detection, and
localization. We conclude by discussing the challenges associated with the use
of 3D CNNs in the medical imaging domain (and the use of deep learning models,
in general) and possible future trends in the field.Comment: 13 pages, 4 figures, 2 table
HyperDense-Net: A hyper-densely connected CNN for multi-modal image segmentation
Recently, dense connections have attracted substantial attention in computer
vision because they facilitate gradient flow and implicit deep supervision
during training. Particularly, DenseNet, which connects each layer to every
other layer in a feed-forward fashion, has shown impressive performances in
natural image classification tasks. We propose HyperDenseNet, a 3D fully
convolutional neural network that extends the definition of dense connectivity
to multi-modal segmentation problems. Each imaging modality has a path, and
dense connections occur not only between the pairs of layers within the same
path, but also between those across different paths. This contrasts with the
existing multi-modal CNN approaches, in which modeling several modalities
relies entirely on a single joint layer (or level of abstraction) for fusion,
typically either at the input or at the output of the network. Therefore, the
proposed network has total freedom to learn more complex combinations between
the modalities, within and in-between all the levels of abstraction, which
increases significantly the learning representation. We report extensive
evaluations over two different and highly competitive multi-modal brain tissue
segmentation challenges, iSEG 2017 and MRBrainS 2013, with the former focusing
on 6-month infant data and the latter on adult images. HyperDenseNet yielded
significant improvements over many state-of-the-art segmentation networks,
ranking at the top on both benchmarks. We further provide a comprehensive
experimental analysis of features re-use, which confirms the importance of
hyper-dense connections in multi-modal representation learning. Our code is
publicly available at https://www.github.com/josedolz/HyperDenseNet.Comment: Paper accepted at IEEE TMI in October 2018. Last version of this
paper updates the reference to the IEEE TMI paper which compares the
submissions to the iSEG 2017 MICCAI Challeng
NiftyNet: a deep-learning platform for medical imaging
Medical image analysis and computer-assisted intervention problems are
increasingly being addressed with deep-learning-based solutions. Established
deep-learning platforms are flexible but do not provide specific functionality
for medical image analysis and adapting them for this application requires
substantial implementation effort. Thus, there has been substantial duplication
of effort and incompatible infrastructure developed across many research
groups. This work presents the open-source NiftyNet platform for deep learning
in medical imaging. The ambition of NiftyNet is to accelerate and simplify the
development of these solutions, and to provide a common mechanism for
disseminating research outputs for the community to use, adapt and build upon.
NiftyNet provides a modular deep-learning pipeline for a range of medical
imaging applications including segmentation, regression, image generation and
representation learning applications. Components of the NiftyNet pipeline
including data loading, data augmentation, network architectures, loss
functions and evaluation metrics are tailored to, and take advantage of, the
idiosyncracies of medical image analysis and computer-assisted intervention.
NiftyNet is built on TensorFlow and supports TensorBoard visualization of 2D
and 3D images and computational graphs by default.
We present 3 illustrative medical image analysis applications built using
NiftyNet: (1) segmentation of multiple abdominal organs from computed
tomography; (2) image regression to predict computed tomography attenuation
maps from brain magnetic resonance images; and (3) generation of simulated
ultrasound images for specified anatomical poses.
NiftyNet enables researchers to rapidly develop and distribute deep learning
solutions for segmentation, regression, image generation and representation
learning applications, or extend the platform to new applications.Comment: Wenqi Li and Eli Gibson contributed equally to this work. M. Jorge
Cardoso and Tom Vercauteren contributed equally to this work. 26 pages, 6
figures; Update includes additional applications, updated author list and
formatting for journal submissio
- …