944 research outputs found
Butterfly Factorization
The paper introduces the butterfly factorization as a data-sparse
approximation for the matrices that satisfy a complementary low-rank property.
The factorization can be constructed efficiently if either fast algorithms for
applying the matrix and its adjoint are available or the entries of the matrix
can be sampled individually. For an matrix, the resulting
factorization is a product of sparse matrices, each with
non-zero entries. Hence, it can be applied rapidly in operations.
Numerical results are provided to demonstrate the effectiveness of the
butterfly factorization and its construction algorithms
The Tensor Networks Anthology: Simulation techniques for many-body quantum lattice systems
We present a compendium of numerical simulation techniques, based on tensor
network methods, aiming to address problems of many-body quantum mechanics on a
classical computer. The core setting of this anthology are lattice problems in
low spatial dimension at finite size, a physical scenario where tensor network
methods, both Density Matrix Renormalization Group and beyond, have long proven
to be winning strategies. Here we explore in detail the numerical frameworks
and methods employed to deal with low-dimension physical setups, from a
computational physics perspective. We focus on symmetries and closed-system
simulations in arbitrary boundary conditions, while discussing the numerical
data structures and linear algebra manipulation routines involved, which form
the core libraries of any tensor network code. At a higher level, we put the
spotlight on loop-free network geometries, discussing their advantages, and
presenting in detail algorithms to simulate low-energy equilibrium states.
Accompanied by discussions of data structures, numerical techniques and
performance, this anthology serves as a programmer's companion, as well as a
self-contained introduction and review of the basic and selected advanced
concepts in tensor networks, including examples of their applications.Comment: 115 pages, 56 figure
Flexible Multi-layer Sparse Approximations of Matrices and Applications
The computational cost of many signal processing and machine learning
techniques is often dominated by the cost of applying certain linear operators
to high-dimensional vectors. This paper introduces an algorithm aimed at
reducing the complexity of applying linear operators in high dimension by
approximately factorizing the corresponding matrix into few sparse factors. The
approach relies on recent advances in non-convex optimization. It is first
explained and analyzed in details and then demonstrated experimentally on
various problems including dictionary learning for image denoising, and the
approximation of large matrices arising in inverse problems
Randomized methods for matrix computations
The purpose of this text is to provide an accessible introduction to a set of
recently developed algorithms for factorizing matrices. These new algorithms
attain high practical speed by reducing the dimensionality of intermediate
computations using randomized projections. The algorithms are particularly
powerful for computing low-rank approximations to very large matrices, but they
can also be used to accelerate algorithms for computing full factorizations of
matrices. A key competitive advantage of the algorithms described is that they
require less communication than traditional deterministic methods
Study of compression techniques for partial differential equation solvers
Partial Differential Equations (PDEs) are widely applied in many branches of science, and solving them efficiently, from a computational point of view, is one of the cornerstones of modern computational science. The finite element (FE) method is a popular numerical technique for calculating approximate solutions to PDEs. A not necessarily complex finite element analysis containing substructures can easily gen-erate enormous quantities of elements that hinder and slow down simulations. Therefore, compression methods are required to decrease the amount of computational effort while retaining the significant dynamics of the problem. In this study, it was decided to apply a purely algebraic approach. Various methods will be included and discussed, ranging from research-level techniques to other apparently unrelated fields like image compression, via the discrete Fourier transform (DFT) and the Wavelet transform or the Singular Value Decomposition (SVD)
Compression of unitary rank--structured matrices to CMV-like shape with an application to polynomial rootfinding
This paper is concerned with the reduction of a unitary matrix U to CMV-like
shape. A Lanczos--type algorithm is presented which carries out the reduction
by computing the block tridiagonal form of the Hermitian part of U, i.e., of
the matrix U+U^H. By elaborating on the Lanczos approach we also propose an
alternative algorithm using elementary matrices which is numerically stable. If
U is rank--structured then the same property holds for its Hermitian part and,
therefore, the block tridiagonalization process can be performed using the
rank--structured matrix technology with reduced complexity. Our interest in the
CMV-like reduction is motivated by the unitary and almost unitary eigenvalue
problem. In this respect, finally, we discuss the application of the CMV-like
reduction for the design of fast companion eigensolvers based on the customary
QR iteration
- …