18 research outputs found
Augmented Block-Arnoldi Recycling CFD Solvers
One of the limitations of recycled GCRO methods is the large amount of
computation required to orthogonalize the basis vectors of the newly generated
Krylov subspace for the approximate solution when combined with those of the
recycle subspace. Recent advancements in low synchronization Gram-Schmidt and
generalized minimal residual algorithms, Swirydowicz et
al.~\cite{2020-swirydowicz-nlawa}, Carson et al. \cite{Carson2022}, and Lund
\cite{Lund2022}, can be incorporated, thereby mitigating the loss of
orthogonality of the basis vectors. An augmented Arnoldi formulation of
recycling leads to a matrix decomposition and the associated algorithm can also
be viewed as a {\it block} Krylov method. Generalizations of both classical and
modified block Gram-Schmidt algorithms have been proposed, Carson et
al.~\cite{Carson2022}. Here, an inverse compact modified Gram-Schmidt
algorithm is applied for the inter-block orthogonalization scheme with a block
lower triangular correction matrix at iteration . When combined with a
weighted (oblique inner product) projection step, the inverse compact
scheme leads to significant (over 10 in certain cases) reductions in
the number of solver iterations per linear system. The weight is also
interpreted in terms of the angle between restart residuals in LGMRES, as
defined by Baker et al.\cite{Baker2005}. In many cases, the recycle subspace
eigen-spectrum can substitute for a preconditioner
Data-driven acceleration of Photonic Simulations
Designing modern photonic devices often involves traversing a large parameter
space via an optimization procedure, gradient based or otherwise, and typically
results in the designer performing electromagnetic simulations of correlated
devices. In this paper, we present an approach to accelerate the Generalized
Minimal Residual (GMRES) algorithm for the solution of frequency-domain
Maxwell's equations using two machine learning models (principal component
analysis and a convolutional neural network) trained on simulations of
correlated devices. These data-driven models are trained to predict a subspace
within which the solution of the frequency-domain Maxwell's equations lie. This
subspace can then be used for augmenting the Krylov subspace generated during
the GMRES iterations. By training the proposed models on a dataset of grating
wavelength-splitting devices, we show an order of magnitude reduction () in the number of GMRES iterations required for solving frequency-domain
Maxwell's equations
A framework for deflated and augmented Krylov subspace methods
We consider deflation and augmentation techniques for accelerating the
convergence of Krylov subspace methods for the solution of nonsingular linear
algebraic systems. Despite some formal similarity, the two techniques are
conceptually different from preconditioning. Deflation (in the sense the term
is used here) "removes" certain parts from the operator making it singular,
while augmentation adds a subspace to the Krylov subspace (often the one that
is generated by the singular operator); in contrast, preconditioning changes
the spectrum of the operator without making it singular. Deflation and
augmentation have been used in a variety of methods and settings. Typically,
deflation is combined with augmentation to compensate for the singularity of
the operator, but both techniques can be applied separately.
We introduce a framework of Krylov subspace methods that satisfy a Galerkin
condition. It includes the families of orthogonal residual (OR) and minimal
residual (MR) methods. We show that in this framework augmentation can be
achieved either explicitly or, equivalently, implicitly by projecting the
residuals appropriately and correcting the approximate solutions in a final
step. We study conditions for a breakdown of the deflated methods, and we show
several possibilities to avoid such breakdowns for the deflated MINRES method.
Numerical experiments illustrate properties of different variants of deflated
MINRES analyzed in this paper.Comment: 24 pages, 3 figure