34 research outputs found
On Learning the Invisible in Photoacoustic Tomography with Flat Directionally Sensitive Detector
In photoacoustic tomography (PAT) with flat sensor, we routinely encounter
two types of limited data. The first is due to using a finite sensor and is
especially perceptible if the region of interest is large relatively to the
sensor or located farther away from the sensor. In this paper, we focus on the
second type caused by a varying sensitivity of the sensor to the incoming
wavefront direction which can be modelled as binary i.e. by a cone of
sensitivity. Such visibility conditions result, in Fourier domain, in a
restriction of both the image and the data to a bowtie, akin to the one
corresponding to the range of the forward operator. The visible ranges, in
image and data domains, are related by the wavefront direction mapping. We
adapt the wedge restricted Curvelet decomposition, we previously proposed for
the representation of the full PAT data, to separate the visible and invisible
wavefronts in the image. We optimally combine fast approximate operators with
tailored deep neural network architectures into efficient learned
reconstruction methods which perform reconstruction of the visible coefficients
and the invisible coefficients are learned from a training set of similar data.Comment: Submitted to SIAM Journal on Imaging Science
Parallel-in-Time Solutions with Random Projection Neural Networks
This paper considers one of the fundamental parallel-in-time methods for the solution of ordinary differential equations, Parareal, and extends it by adopting a neural network as a coarse propagator. We provide a theoretical analysis of the convergence properties of the proposed algorithm and show its effectiveness for several examples, including Lorenz and Burgers' equations. In our numerical simulations, we further specialize the underpinning neural architecture to Random Projection Neural Networks (RPNNs), a 2-layer neural network where the first layer weights are drawn at random rather than optimized. This restriction substantially increases the efficiency of fitting RPNN's weights in comparison to a standard feedforward network without negatively impacting the accuracy, as demonstrated in the SIR system example
Learned Interferometric Imaging for the SPIDER Instrument
The Segmented Planar Imaging Detector for Electro-Optical Reconnaissance
(SPIDER) is an optical interferometric imaging device that aims to offer an
alternative to the large space telescope designs of today with reduced size,
weight and power consumption. This is achieved through interferometric imaging.
State-of-the-art methods for reconstructing images from interferometric
measurements adopt proximal optimization techniques, which are computationally
expensive and require handcrafted priors. In this work we present two
data-driven approaches for reconstructing images from measurements made by the
SPIDER instrument. These approaches use deep learning to learn prior
information from training data, increasing the reconstruction quality, and
significantly reducing the computation time required to recover images by
orders of magnitude. Reconstruction time is reduced to
milliseconds, opening up the possibility of real-time imaging with SPIDER for
the first time. Furthermore, we show that these methods can also be applied in
domains where training data is scarce, such as astronomical imaging, by
leveraging transfer learning from domains where plenty of training data are
available.Comment: 21 pages, 14 figure
On the Adjoint Operator in Photoacoustic Tomography
Photoacoustic Tomography (PAT) is an emerging biomedical "imaging from
coupled physics" technique, in which the image contrast is due to optical
absorption, but the information is carried to the surface of the tissue as
ultrasound pulses. Many algorithms and formulae for PAT image reconstruction
have been proposed for the case when a complete data set is available. In many
practical imaging scenarios, however, it is not possible to obtain the full
data, or the data may be sub-sampled for faster data acquisition. In such
cases, image reconstruction algorithms that can incorporate prior knowledge to
ameliorate the loss of data are required. Hence, recently there has been an
increased interest in using variational image reconstruction. A crucial
ingredient for the application of these techniques is the adjoint of the PAT
forward operator, which is described in this article from physical, theoretical
and numerical perspectives. First, a simple mathematical derivation of the
adjoint of the PAT forward operator in the continuous framework is presented.
Then, an efficient numerical implementation of the adjoint using a k-space time
domain wave propagation model is described and illustrated in the context of
variational PAT image reconstruction, on both 2D and 3D examples including
inhomogeneous sound speed. The principal advantage of this analytical adjoint
over an algebraic adjoint (obtained by taking the direct adjoint of the
particular numerical forward scheme used) is that it can be implemented using
currently available fast wave propagation solvers.Comment: submitted to "Inverse Problems
Restarting projection methods for rational eigenproblems arising in fluidâsolid vibrations
For nonlinear eigenvalue problems T(λ)x = 0 satisfying a minmax characterization of its eigenvalues iterative projection methods combined with safeguarded iteration are suitable for computing all eigenvalues in a given interval. Such methods hit their limitations if a large number of eigenvalues is required. In this paper we discuss restart procedures which are able to cope with this problem, and we evaluate them for a rational eigenvalue problem governing vibrations of a fluidâsolid structure.
First Published Online:Â 14 Oct 201
Choose your path wisely: gradient descent in a Bregman distance framework
We propose an extension of a special form of gradient descent --- in the
literature known as linearised Bregman iteration -- to a larger class of
non-convex functions. We replace the classical (squared) two norm metric in the
gradient descent setting with a generalised Bregman distance, based on a
proper, convex and lower semi-continuous function. The algorithm's global
convergence is proven for functions that satisfy the Kurdyka-\L ojasiewicz
property. Examples illustrate that features of different scale are being
introduced throughout the iteration, transitioning from coarse to fine. This
coarse-to-fine approach with respect to scale allows to recover solutions of
non-convex optimisation problems that are superior to those obtained with
conventional gradient descent, or even projected and proximal gradient descent.
The effectiveness of the linearised Bregman iteration in combination with early
stopping is illustrated for the applications of parallel magnetic resonance
imaging, blind deconvolution as well as image classification with neural
networks