298 research outputs found
Lecture 13: A low-rank factorization framework for building scalable algebraic solvers and preconditioners
Factorization based preconditioning algorithms, most notably incomplete LU (ILU) factorization, have been shown to be robust and applicable to wide ranges of problems. However, traditional ILU algorithms are not amenable to scalable implementation. In recent years, we have seen a lot of investigations using low-rank compression techniques to build approximate factorizations.A key to achieving lower complexity is the use of hierarchical matrix algebra, stemming from the H-matrix research. In addition, the multilevel algorithm paradigm provides a good vehicle for a scalable implementation. The goal of this lecture is to give an overview of the various hierarchical matrix formats, such as hierarchically semi-separable matrix (HSS), hierarchically off-diagonal low-rank matrix (HODLR) and Butterfly matrix, and explain the algorithm differences and approximation quality. We will illustrate many practical issues of these algorithms using our parallel libraries STRUMPACK and ButterflyPACK, and demonstrate their effectiveness and scalability while solving the very challenging problems, such as high frequency wave equations
Randomized Local Model Order Reduction
In this paper we propose local approximation spaces for localized model order
reduction procedures such as domain decomposition and multiscale methods. Those
spaces are constructed from local solutions of the partial differential
equation (PDE) with random boundary conditions, yield an approximation that
converges provably at a nearly optimal rate, and can be generated at close to
optimal computational complexity. In many localized model order reduction
approaches like the generalized finite element method, static condensation
procedures, and the multiscale finite element method local approximation spaces
can be constructed by approximating the range of a suitably defined transfer
operator that acts on the space of local solutions of the PDE. Optimal local
approximation spaces that yield in general an exponentially convergent
approximation are given by the left singular vectors of this transfer operator
[I. Babu\v{s}ka and R. Lipton 2011, K. Smetana and A. T. Patera 2016]. However,
the direct calculation of these singular vectors is computationally very
expensive. In this paper, we propose an adaptive randomized algorithm based on
methods from randomized linear algebra [N. Halko et al. 2011], which constructs
a local reduced space approximating the range of the transfer operator and thus
the optimal local approximation spaces. The adaptive algorithm relies on a
probabilistic a posteriori error estimator for which we prove that it is both
efficient and reliable with high probability. Several numerical experiments
confirm the theoretical findings.Comment: 31 pages, 14 figures, 1 table, 1 algorith
Solving high-dimensional Fokker-Planck equation with functional hierarchical tensor
This work is concerned with solving high-dimensional Fokker-Planck equations
with the novel perspective that solving the PDE can be reduced to independent
instances of density estimation tasks based on the trajectories sampled from
its associated particle dynamics. With this approach, one sidesteps error
accumulation occurring from integrating the PDE dynamics on a parameterized
function class. This approach significantly simplifies deployment, as one is
free of the challenges of implementing loss terms based on the differential
equation. In particular, we introduce a novel class of high-dimensional
functions called the functional hierarchical tensor (FHT). The FHT ansatz
leverages a hierarchical low-rank structure, offering the advantage of linearly
scalable runtime and memory complexity relative to the dimension count. We
introduce a sketching-based technique that performs density estimation over
particles simulated from the particle dynamics associated with the equation,
thereby obtaining a representation of the Fokker-Planck solution in terms of
our ansatz. We apply the proposed approach successfully to three challenging
time-dependent Ginzburg-Landau models with hundreds of variables
Lecture 13: A low-rank factorization framework for building scalable algebraic solvers and preconditioners
Factorization based preconditioning algorithms, most notably incomplete LU (ILU) factorization, have been shown to be robust and applicable to wide ranges of problems. However, traditional ILU algorithms are not amenable to scalable implementation. In recent years, we have seen a lot of investigations using low-rank compression techniques to build approximate factorizations.A key to achieving lower complexity is the use of hierarchical matrix algebra, stemming from the H-matrix research. In addition, the multilevel algorithm paradigm provides a good vehicle for a scalable implementation. The goal of this lecture is to give an overview of the various hierarchical matrix formats, such as hierarchically semi-separable matrix (HSS), hierarchically off-diagonal low-rank matrix (HODLR) and Butterfly matrix, and explain the algorithm differences and approximation quality. We will illustrate many practical issues of these algorithms using our parallel libraries STRUMPACK and ButterflyPACK, and demonstrate their effectiveness and scalability while solving the very challenging problems, such as high frequency wave equations
Lecture 02: Tile Low-rank Methods and Applications (w/review)
As simulation and analytics enter the exascale era, numerical algorithms, particularly implicit solvers that couple vast numbers of degrees of freedom, must span a widening gap between ambitious applications and austere architectures to support them. We present fifteen universals for researchers in scalable solvers: imperatives from computer architecture that scalable solvers must respect, strategies towards achieving them that are currently well established, and additional strategies currently being developed for an effective and efficient exascale software ecosystem. We consider recent generalizations of what it means to “solve” a computational problem, which suggest that we have often been “oversolving” them at the smaller scales of the past because we could afford to do so. We present innovations that allow to approach lin-log complexity in storage and operation count in many important algorithmic kernels and thus create an opportunity for full applications with optimal scalability
Dictionary-based model reduction for state estimation
We consider the problem of state estimation from linear measurements,
where the state to recover is an element of the manifold of
solutions of a parameter-dependent equation. The state is estimated using a
prior knowledge on coming from model order reduction. Variational
approaches based on linear approximation of , such as PBDW, yields
a recovery error limited by the Kolmogorov -width of . To
overcome this issue, piecewise-affine approximations of have also
be considered, that consist in using a library of linear spaces among which one
is selected by minimizing some distance to . In this paper, we
propose a state estimation method relying on dictionary-based model reduction,
where a space is selected from a library generated by a dictionary of
snapshots, using a distance to the manifold. The selection is performed among a
set of candidate spaces obtained from the path of a -regularized
least-squares problem. Then, in the framework of parameter-dependent operator
equations (or PDEs) with affine parameterizations, we provide an efficient
offline-online decomposition based on randomized linear algebra, that ensures
efficient and stable computations while preserving theoretical guarantees.Comment: 19 pages, 5 figure
- …