1,036 research outputs found
A Fast Algorithm for Parabolic PDE-based Inverse Problems Based on Laplace Transforms and Flexible Krylov Solvers
We consider the problem of estimating parameters in large-scale weakly
nonlinear inverse problems for which the underlying governing equations is a
linear, time-dependent, parabolic partial differential equation. A major
challenge in solving these inverse problems using Newton-type methods is the
computational cost associated with solving the forward problem and with
repeated construction of the Jacobian, which represents the sensitivity of the
measurements to the unknown parameters. Forming the Jacobian can be
prohibitively expensive because it requires repeated solutions of the forward
and adjoint time-dependent parabolic partial differential equations
corresponding to multiple sources and receivers. We propose an efficient method
based on a Laplace transform-based exponential time integrator combined with a
flexible Krylov subspace approach to solve the resulting shifted systems of
equations efficiently. Our proposed solver speeds up the computation of the
forward and adjoint problems, thus yielding significant speedup in total
inversion time. We consider an application from Transient Hydraulic Tomography
(THT), which is an imaging technique to estimate hydraulic parameters related
to the subsurface from pressure measurements obtained by a series of pumping
tests. The algorithms discussed are applied to a synthetic example taken from
THT to demonstrate the resulting computational gains of this proposed method
Randomized Riemannian Preconditioning for Orthogonality Constrained Problems
Optimization problems with (generalized) orthogonality constraints are
prevalent across science and engineering. For example, in computational science
they arise in the symmetric (generalized) eigenvalue problem, in nonlinear
eigenvalue problems, and in electronic structures computations, to name a few
problems. In statistics and machine learning, they arise, for example, in
canonical correlation analysis and in linear discriminant analysis. In this
article, we consider using randomized preconditioning in the context of
optimization problems with generalized orthogonality constraints. Our proposed
algorithms are based on Riemannian optimization on the generalized Stiefel
manifold equipped with a non-standard preconditioned geometry, which
necessitates development of the geometric components necessary for developing
algorithms based on this approach. Furthermore, we perform asymptotic
convergence analysis of the preconditioned algorithms which help to
characterize the quality of a given preconditioner using second-order
information. Finally, for the problems of canonical correlation analysis and
linear discriminant analysis, we develop randomized preconditioners along with
corresponding bounds on the relevant condition number
- …