415 research outputs found
A class of nonsymmetric preconditioners for saddle point problems
For iterative solution of saddle point problems, a nonsymmetric preconditioning is studied which, with respect to the upper-left block of the system matrix, can be seen as a variant of SSOR. An idealized situation where the SSOR is taken with respect to the skew-symmetric part plus the diagonal part of the upper-left block is analyzed in detail. Since action of the preconditioner involves solution of a Schur complement system, an inexact form of the preconditioner can be of interest. This results in an inner-outer iterative process. Numerical experiments with solution of linearized Navier-Stokes equations demonstrate efficiency of the new preconditioner, especially when the left-upper block is far from symmetric
Updating constraint preconditioners for KKT systems in quadratic programming via low-rank corrections
This work focuses on the iterative solution of sequences of KKT linear
systems arising in interior point methods applied to large convex quadratic
programming problems. This task is the computational core of the interior point
procedure and an efficient preconditioning strategy is crucial for the
efficiency of the overall method. Constraint preconditioners are very effective
in this context; nevertheless, their computation may be very expensive for
large-scale problems, and resorting to approximations of them may be
convenient. Here we propose a procedure for building inexact constraint
preconditioners by updating a "seed" constraint preconditioner computed for a
KKT matrix at a previous interior point iteration. These updates are obtained
through low-rank corrections of the Schur complement of the (1,1) block of the
seed preconditioner. The updated preconditioners are analyzed both
theoretically and computationally. The results obtained show that our updating
procedure, coupled with an adaptive strategy for determining whether to
reinitialize or update the preconditioner, can enhance the performance of
interior point methods on large problems.Comment: 22 page
A preconditioned MINRES method for nonsymmetric Toeplitz matrices
Circulant preconditioning for symmetric Toeplitz linear systems is well established; theoretical guarantees of fast convergence for the conjugate gradient method are descriptive of the convergence seen in computations. This has led to robust and highly efficient solvers based on use of the fast Fourier transform exactly as originally envisaged in [G. Strang, Stud. Appl. Math., 74 (1986), pp. 171--176]. For nonsymmetric systems, the lack of generally descriptive convergence theory for most iterative methods of Krylov type has provided a barrier to such a comprehensive guarantee, though several methods have been proposed and some analysis of performance with the normal equations is available. In this paper, by the simple device of reordering, we rigorously establish a circulant preconditioned short recurrence Krylov subspace iterative method of minimum residual type for nonsymmetric (and possibly highly nonnormal) Toeplitz systems. Convergence estimates similar to those in the symmetric case are established
Approximation of the scattering amplitude
The simultaneous solution of Ax=b and ATy=g is required in a number of situations. Darmofal and Lu have proposed a method based on the Quasi-Minimal residual algorithm (QMR). We will introduce a technique for the same purpose based on the LSQR method and show how its performance can be improved when using the Generalized LSQR method. We further show how preconditioners can be introduced to enhance the speed of convergence and discuss different preconditioners that can be used. The scattering amplitude gTx, a widely used quantity in signal processing for example, has a close connection to the above problem since x represents the solution of the forward problem and g is the right hand side of the adjoint system. We show how this quantity can be efficiently approximated using Gauss quadrature and introduce a Block-Lanczos process that approximates the scattering amplitude and which can also be used with preconditioners
Absolute value preconditioning for symmetric indefinite linear systems
We introduce a novel strategy for constructing symmetric positive definite
(SPD) preconditioners for linear systems with symmetric indefinite matrices.
The strategy, called absolute value preconditioning, is motivated by the
observation that the preconditioned minimal residual method with the inverse of
the absolute value of the matrix as a preconditioner converges to the exact
solution of the system in at most two steps. Neither the exact absolute value
of the matrix nor its exact inverse are computationally feasible to construct
in general. However, we provide a practical example of an SPD preconditioner
that is based on the suggested approach. In this example we consider a model
problem with a shifted discrete negative Laplacian, and suggest a geometric
multigrid (MG) preconditioner, where the inverse of the matrix absolute value
appears only on the coarse grid, while operations on finer grids are based on
the Laplacian. Our numerical tests demonstrate practical effectiveness of the
new MG preconditioner, which leads to a robust iterative scheme with minimalist
memory requirements
Multilevel Solvers for Unstructured Surface Meshes
Parameterization of unstructured surface meshes is of fundamental importance in many applications of digital geometry processing. Such parameterization approaches give rise to large and exceedingly ill-conditioned systems which are difficult or impossible to solve without the use of sophisticated multilevel preconditioning strategies. Since the underlying meshes are very fine to begin with, such multilevel preconditioners require mesh coarsening to build an appropriate hierarchy. In this paper we consider several strategies for the construction of hierarchies using ideas from mesh simplification algorithms used in the computer graphics literature. We introduce two novel hierarchy construction schemes and demonstrate their superior performance when used in conjunction with a multigrid preconditioner
A biconjugate gradient type algorithm on massively parallel architectures
The biconjugate gradient (BCG) method is the natural generalization of the classical conjugate gradient algorithm for Hermitian positive definite matrices to general non-Hermitian linear systems. Unfortunately, the original BCG algorithm is susceptible to possible breakdowns and numerical instabilities. Recently, Freund and Nachtigal have proposed a novel BCG type approach, the quasi-minimal residual method (QMR), which overcomes the problems of BCG. Here, an implementation is presented of QMR based on an s-step version of the nonsymmetric look-ahead Lanczos algorithm. The main feature of the s-step Lanczos algorithm is that, in general, all inner products, except for one, can be computed in parallel at the end of each block; this is unlike the other standard Lanczos process where inner products are generated sequentially. The resulting implementation of QMR is particularly attractive on massively parallel SIMD architectures, such as the Connection Machine
A Combined Preconditioning Strategy for Nonsymmetric Systems
We present and analyze a class of nonsymmetric preconditioners within a
normal (weighted least-squares) matrix form for use in GMRES to solve
nonsymmetric matrix problems that typically arise in finite element
discretizations. An example of the additive Schwarz method applied to
nonsymmetric but definite matrices is presented for which the abstract
assumptions are verified. A variable preconditioner, combining the original
nonsymmetric one and a weighted least-squares version of it, is shown to be
convergent and provides a viable strategy for using nonsymmetric
preconditioners in practice. Numerical results are included to assess the
theory and the performance of the proposed preconditioners.Comment: 26 pages, 3 figure
A framework for deflated and augmented Krylov subspace methods
We consider deflation and augmentation techniques for accelerating the
convergence of Krylov subspace methods for the solution of nonsingular linear
algebraic systems. Despite some formal similarity, the two techniques are
conceptually different from preconditioning. Deflation (in the sense the term
is used here) "removes" certain parts from the operator making it singular,
while augmentation adds a subspace to the Krylov subspace (often the one that
is generated by the singular operator); in contrast, preconditioning changes
the spectrum of the operator without making it singular. Deflation and
augmentation have been used in a variety of methods and settings. Typically,
deflation is combined with augmentation to compensate for the singularity of
the operator, but both techniques can be applied separately.
We introduce a framework of Krylov subspace methods that satisfy a Galerkin
condition. It includes the families of orthogonal residual (OR) and minimal
residual (MR) methods. We show that in this framework augmentation can be
achieved either explicitly or, equivalently, implicitly by projecting the
residuals appropriately and correcting the approximate solutions in a final
step. We study conditions for a breakdown of the deflated methods, and we show
several possibilities to avoid such breakdowns for the deflated MINRES method.
Numerical experiments illustrate properties of different variants of deflated
MINRES analyzed in this paper.Comment: 24 pages, 3 figure
- …