5 research outputs found
Learning with Known Operators reduces Maximum Training Error Bounds
We describe an approach for incorporating prior knowledge into machine
learning algorithms. We aim at applications in physics and signal processing in
which we know that certain operations must be embedded into the algorithm. Any
operation that allows computation of a gradient or sub-gradient towards its
inputs is suited for our framework. We derive a maximal error bound for deep
nets that demonstrates that inclusion of prior knowledge results in its
reduction. Furthermore, we also show experimentally that known operators reduce
the number of free parameters. We apply this approach to various tasks ranging
from CT image reconstruction over vessel segmentation to the derivation of
previously unknown imaging algorithms. As such the concept is widely applicable
for many researchers in physics, imaging, and signal processing. We assume that
our analysis will support further investigation of known operators in other
fields of physics, imaging, and signal processing.Comment: Paper conditionally accepted in Nature Machine Intelligenc
Field of View Extension in Computed Tomography Using Deep Learning Prior
In computed tomography (CT), data truncation is a common problem. Images
reconstructed by the standard filtered back-projection algorithm from truncated
data suffer from cupping artifacts inside the field-of-view (FOV), while
anatomical structures are severely distorted or missing outside the FOV. Deep
learning, particularly the U-Net, has been applied to extend the FOV as a
post-processing method. Since image-to-image prediction neglects the data
fidelity to measured projection data, incorrect structures, even inside the
FOV, might be reconstructed by such an approach. Therefore, generating
reconstructed images directly from a post-processing neural network is
inadequate. In this work, we propose a data consistent reconstruction method,
which utilizes deep learning reconstruction as prior for extrapolating
truncated projections and a conventional iterative reconstruction to constrain
the reconstruction consistent to measured raw data. Its efficacy is
demonstrated in our study, achieving small average root-mean-square error of 24
HU inside the FOV and a high structure similarity index of 0.993 for the whole
body area on a test patient's CT data.Comment: Submitted to Bildverarbeitung fuer die Medizin 202
RinQ Fingerprinting: Recurrence-informed Quantile Networks for Magnetic Resonance Fingerprinting
Recently, Magnetic Resonance Fingerprinting (MRF) was proposed as a
quantitative imaging technique for the simultaneous acquisition of tissue
parameters such as relaxation times and . Although the acquisition
is highly accelerated, the state-of-the-art reconstruction suffers from long
computation times: Template matching methods are used to find the most similar
signal to the measured one by comparing it to pre-simulated signals of possible
parameter combinations in a discretized dictionary. Deep learning approaches
can overcome this limitation, by providing the direct mapping from the measured
signal to the underlying parameters by one forward pass through a network. In
this work, we propose a Recurrent Neural Network (RNN) architecture in
combination with a novel quantile layer. RNNs are well suited for the
processing of time-dependent signals and the quantile layer helps to overcome
the noisy outliers by considering the spatial neighbors of the signal. We
evaluate our approach using in-vivo data from multiple brain slices and several
volunteers, running various experiments. We show that the RNN approach with
small patches of complex-valued input signals in combination with a quantile
layer outperforms other architectures, e.g. previously proposed CNNs for the
MRF reconstruction reducing the error in and by more than 80%.Comment: Accepted for MICCAI 201
Data Consistent Artifact Reduction for Limited Angle Tomography with Deep Learning Prior
Robustness of deep learning methods for limited angle tomography is
challenged by two major factors: a) due to insufficient training data the
network may not generalize well to unseen data; b) deep learning methods are
sensitive to noise. Thus, generating reconstructed images directly from a
neural network appears inadequate. We propose to constrain the reconstructed
images to be consistent with the measured projection data, while the unmeasured
information is complemented by learning based methods. For this purpose, a data
consistent artifact reduction (DCAR) method is introduced: First, a prior image
is generated from an initial limited angle reconstruction via deep learning as
a substitute for missing information. Afterwards, a conventional iterative
reconstruction algorithm is applied, integrating the data consistency in the
measured angular range and the prior information in the missing angular range.
This ensures data integrity in the measured area, while inaccuracies
incorporated by the deep learning prior lie only in areas where no information
is acquired. The proposed DCAR method achieves significant image quality
improvement: for 120-degree cone-beam limited angle tomography more than 10%
RMSE reduction in noise-free case and more than 24% RMSE reduction in noisy
case compared with a state-of-the-art U-Net based method.Comment: Accepted by MICCAI MLMIR worksho
Data Consistent CT Reconstruction from Insufficient Data with Learned Prior Images
Image reconstruction from insufficient data is common in computed tomography
(CT), e.g., image reconstruction from truncated data, limited-angle data and
sparse-view data. Deep learning has achieved impressive results in this field.
However, the robustness of deep learning methods is still a concern for
clinical applications due to the following two challenges: a) With limited
access to sufficient training data, a learned deep learning model may not
generalize well to unseen data; b) Deep learning models are sensitive to noise.
Therefore, the quality of images processed by neural networks only may be
inadequate. In this work, we investigate the robustness of deep learning in CT
image reconstruction by showing false negative and false positive lesion cases.
Since learning-based images with incorrect structures are likely not consistent
with measured projection data, we propose a data consistent reconstruction
(DCR) method to improve their image quality, which combines the advantages of
compressed sensing and deep learning: First, a prior image is generated by deep
learning. Afterwards, unmeasured projection data are inpainted by forward
projection of the prior image. Finally, iterative reconstruction with
reweighted total variation regularization is applied, integrating data
consistency for measured data and learned prior information for missing data.
The efficacy of the proposed method is demonstrated in cone-beam CT with
truncated data, limited-angle data and sparse-view data, respectively. For
example, for truncated data, DCR achieves a mean root-mean-square error of 24
HU and a mean structure similarity index of 0.999 inside the field-of-view for
different patients in the noisy case, while the state-of-the-art U-Net method
achieves 55 HU and 0.995 respectively for these two metrics.Comment: 10 pages, 9 figure