2,282 research outputs found
Intraoperative Organ Motion Models with an Ensemble of Conditional Generative Adversarial Networks
In this paper, we describe how a patient-specific, ultrasound-probe-induced
prostate motion model can be directly generated from a single preoperative MR
image. Our motion model allows for sampling from the conditional distribution
of dense displacement fields, is encoded by a generative neural network
conditioned on a medical image, and accepts random noise as additional input.
The generative network is trained by a minimax optimisation with a second
discriminative neural network, tasked to distinguish generated samples from
training motion data. In this work, we propose that 1) jointly optimising a
third conditioning neural network that pre-processes the input image, can
effectively extract patient-specific features for conditioning; and 2)
combining multiple generative models trained separately with heuristically
pre-disjointed training data sets can adequately mitigate the problem of mode
collapse. Trained with diagnostic T2-weighted MR images from 143 real patients
and 73,216 3D dense displacement fields from finite element simulations of
intraoperative prostate motion due to transrectal ultrasound probe pressure,
the proposed models produced physically-plausible patient-specific motion of
prostate glands. The ability to capture biomechanically simulated motion was
evaluated using two errors representing generalisability and specificity of the
model. The median values, calculated from a 10-fold cross-validation, were
2.8+/-0.3 mm and 1.7+/-0.1 mm, respectively. We conclude that the introduced
approach demonstrates the feasibility of applying state-of-the-art machine
learning algorithms to generate organ motion models from patient images, and
shows significant promise for future research.Comment: Accepted to MICCAI 201
Adversarial Deformation Regularization for Training Image Registration Neural Networks
We describe an adversarial learning approach to constrain convolutional
neural network training for image registration, replacing heuristic smoothness
measures of displacement fields often used in these tasks. Using
minimally-invasive prostate cancer intervention as an example application, we
demonstrate the feasibility of utilizing biomechanical simulations to
regularize a weakly-supervised anatomical-label-driven registration network for
aligning pre-procedural magnetic resonance (MR) and 3D intra-procedural
transrectal ultrasound (TRUS) images. A discriminator network is optimized to
distinguish the registration-predicted displacement fields from the motion data
simulated by finite element analysis. During training, the registration network
simultaneously aims to maximize similarity between anatomical labels that
drives image alignment and to minimize an adversarial generator loss that
measures divergence between the predicted- and simulated deformation. The
end-to-end trained network enables efficient and fully-automated registration
that only requires an MR and TRUS image pair as input, without anatomical
labels or simulated data during inference. 108 pairs of labelled MR and TRUS
images from 76 prostate cancer patients and 71,500 nonlinear finite-element
simulations from 143 different patients were used for this study. We show that,
with only gland segmentation as training labels, the proposed method can help
predict physically plausible deformation without any other smoothness penalty.
Based on cross-validation experiments using 834 pairs of independent validation
landmarks, the proposed adversarial-regularized registration achieved a target
registration error of 6.3 mm that is significantly lower than those from
several other regularization methods.Comment: Accepted to MICCAI 201
NiftyNet: a deep-learning platform for medical imaging
Medical image analysis and computer-assisted intervention problems are
increasingly being addressed with deep-learning-based solutions. Established
deep-learning platforms are flexible but do not provide specific functionality
for medical image analysis and adapting them for this application requires
substantial implementation effort. Thus, there has been substantial duplication
of effort and incompatible infrastructure developed across many research
groups. This work presents the open-source NiftyNet platform for deep learning
in medical imaging. The ambition of NiftyNet is to accelerate and simplify the
development of these solutions, and to provide a common mechanism for
disseminating research outputs for the community to use, adapt and build upon.
NiftyNet provides a modular deep-learning pipeline for a range of medical
imaging applications including segmentation, regression, image generation and
representation learning applications. Components of the NiftyNet pipeline
including data loading, data augmentation, network architectures, loss
functions and evaluation metrics are tailored to, and take advantage of, the
idiosyncracies of medical image analysis and computer-assisted intervention.
NiftyNet is built on TensorFlow and supports TensorBoard visualization of 2D
and 3D images and computational graphs by default.
We present 3 illustrative medical image analysis applications built using
NiftyNet: (1) segmentation of multiple abdominal organs from computed
tomography; (2) image regression to predict computed tomography attenuation
maps from brain magnetic resonance images; and (3) generation of simulated
ultrasound images for specified anatomical poses.
NiftyNet enables researchers to rapidly develop and distribute deep learning
solutions for segmentation, regression, image generation and representation
learning applications, or extend the platform to new applications.Comment: Wenqi Li and Eli Gibson contributed equally to this work. M. Jorge
Cardoso and Tom Vercauteren contributed equally to this work. 26 pages, 6
figures; Update includes additional applications, updated author list and
formatting for journal submissio
Label-driven weakly-supervised learning for multimodal deformable image registration
Spatially aligning medical images from different modalities remains a
challenging task, especially for intraoperative applications that require fast
and robust algorithms. We propose a weakly-supervised, label-driven formulation
for learning 3D voxel correspondence from higher-level label correspondence,
thereby bypassing classical intensity-based image similarity measures. During
training, a convolutional neural network is optimised by outputting a dense
displacement field (DDF) that warps a set of available anatomical labels from
the moving image to match their corresponding counterparts in the fixed image.
These label pairs, including solid organs, ducts, vessels, point landmarks and
other ad hoc structures, are only required at training time and can be
spatially aligned by minimising a cross-entropy function of the warped moving
label and the fixed label. During inference, the trained network takes a new
image pair to predict an optimal DDF, resulting in a fully-automatic,
label-free, real-time and deformable registration. For interventional
applications where large global transformation prevails, we also propose a
neural network architecture to jointly optimise the global- and local
displacements. Experiment results are presented based on cross-validating
registrations of 111 pairs of T2-weighted magnetic resonance images and 3D
transrectal ultrasound images from prostate cancer patients with a total of
over 4000 anatomical labels, yielding a median target registration error of 4.2
mm on landmark centroids and a median Dice of 0.88 on prostate glands.Comment: Accepted to ISBI 201
Freehand Ultrasound Image Simulation with Spatially-Conditioned Generative Adversarial Networks
Sonography synthesis has a wide range of applications, including medical
procedure simulation, clinical training and multimodality image registration.
In this paper, we propose a machine learning approach to simulate ultrasound
images at given 3D spatial locations (relative to the patient anatomy), based
on conditional generative adversarial networks (GANs). In particular, we
introduce a novel neural network architecture that can sample anatomically
accurate images conditionally on spatial position of the (real or mock)
freehand ultrasound probe. To ensure an effective and efficient spatial
information assimilation, the proposed spatially-conditioned GANs take
calibrated pixel coordinates in global physical space as conditioning input,
and utilise residual network units and shortcuts of conditioning data in the
GANs' discriminator and generator, respectively. Using optically tracked B-mode
ultrasound images, acquired by an experienced sonographer on a fetus phantom,
we demonstrate the feasibility of the proposed method by two sets of
quantitative results: distances were calculated between corresponding
anatomical landmarks identified in the held-out ultrasound images and the
simulated data at the same locations unseen to the networks; a usability study
was carried out to distinguish the simulated data from the real images. In
summary, we present what we believe are state-of-the-art visually realistic
ultrasound images, simulated by the proposed GAN architecture that is stable to
train and capable of generating plausibly diverse image samples.Comment: Accepted to MICCAI RAMBO 201
Privileged Anatomical and Protocol Discrimination in Trackerless 3D Ultrasound Reconstruction
Three-dimensional (3D) freehand ultrasound (US) reconstruction without using
any additional external tracking device has seen recent advances with deep
neural networks (DNNs). In this paper, we first investigated two identified
contributing factors of the learned inter-frame correlation that enable the
DNN-based reconstruction: anatomy and protocol. We propose to incorporate the
ability to represent these two factors - readily available during training - as
the privileged information to improve existing DNN-based methods. This is
implemented in a new multi-task method, where the anatomical and protocol
discrimination are used as auxiliary tasks. We further develop a differentiable
network architecture to optimise the branching location of these auxiliary
tasks, which controls the ratio between shared and task-specific network
parameters, for maximising the benefits from the two auxiliary tasks.
Experimental results, on a dataset with 38 forearms of 19 volunteers acquired
with 6 different scanning protocols, show that 1) both anatomical and protocol
variances are enabling factors for DNN-based US reconstruction; 2) learning how
to discriminate different subjects (anatomical variance) and predefined types
of scanning paths (protocol variance) both significantly improve frame
prediction accuracy, volume reconstruction overlap, accumulated tracking error
and final drift, using the proposed algorithm.Comment: Accepted to Advances in Simplifying Medical UltraSound (ASMUS)
workshop at MICCAI 202
Trackerless freehand ultrasound with sequence modelling and auxiliary transformation over past and future frames
Three-dimensional (3D) freehand ultrasound (US) reconstruction without a
tracker can be advantageous over its two-dimensional or tracked counterparts in
many clinical applications. In this paper, we propose to estimate 3D spatial
transformation between US frames from both past and future 2D images, using
feed-forward and recurrent neural networks (RNNs). With the temporally
available frames, a further multi-task learning algorithm is proposed to
utilise a large number of auxiliary transformation-predicting tasks between
them. Using more than 40,000 US frames acquired from 228 scans on 38 forearms
of 19 volunteers in a volunteer study, the hold-out test performance is
quantified by frame prediction accuracy, volume reconstruction overlap,
accumulated tracking error and final drift, based on ground-truth from an
optical tracker. The results show the importance of modelling the
temporal-spatially correlated input frames as well as output transformations,
with further improvement owing to additional past and/or future frames. The
best performing model was associated with predicting transformation between
moderately-spaced frames, with an interval of less than ten frames at 20 frames
per second (fps). Little benefit was observed by adding frames more than one
second away from the predicted transformation, with or without LSTM-based RNNs.
Interestingly, with the proposed approach, explicit within-sequence loss that
encourages consistency in composing transformations or minimises accumulated
error may no longer be required. The implementation code and volunteer data
will be made publicly available ensuring reproducibility and further research.Comment: 10 pages, 4 figures, Paper submitted to IEEE International Symposium
on Biomedical Imaging (ISBI
Bi-parametric prostate MR image synthesis using pathology and sequence-conditioned stable diffusion
We propose an image synthesis mechanism for multi-sequence prostate MR images conditioned on text, to control lesion presence and sequence, as well as to generate paired
bi-parametric images conditioned on images e.g. for generating diffusion-weighted MR
from T2-weighted MR for paired data, which are two challenging tasks in pathological image synthesis. Our proposed mechanism utilises and builds upon the recent stable diffusion
model by proposing image-based conditioning for paired data generation. We validate our
method using 2D image slices from real suspected prostate cancer patients. The realism of
the synthesised images is validated by means of a blind expert evaluation for identifying
real versus fake images, where a radiologist with 4 years experience reading urological MR
only achieves 59.4% accuracy across all tested sequences (where chance is 50%). For the
first time, we evaluate the realism of the generated pathology by blind expert identification
of the presence of suspected lesions, where we find that the clinician performs similarly
for both real and synthesised images, with a 2.9 percentage point difference in lesion identification accuracy between real and synthesised images, demonstrating the potentials in
radiological training purposes. Furthermore, we also show that a machine learning model,
trained for lesion identification, shows better performance (76.2% vs 70.4%, statistically
significant improvement) when trained with real data augmented by synthesised data as
opposed to training with only real images, demonstrating usefulness for model training
- …