944 research outputs found

    Butterfly Factorization

    Full text link
    The paper introduces the butterfly factorization as a data-sparse approximation for the matrices that satisfy a complementary low-rank property. The factorization can be constructed efficiently if either fast algorithms for applying the matrix and its adjoint are available or the entries of the matrix can be sampled individually. For an N×NN \times N matrix, the resulting factorization is a product of O(logN)O(\log N) sparse matrices, each with O(N)O(N) non-zero entries. Hence, it can be applied rapidly in O(NlogN)O(N\log N) operations. Numerical results are provided to demonstrate the effectiveness of the butterfly factorization and its construction algorithms

    The Tensor Networks Anthology: Simulation techniques for many-body quantum lattice systems

    Full text link
    We present a compendium of numerical simulation techniques, based on tensor network methods, aiming to address problems of many-body quantum mechanics on a classical computer. The core setting of this anthology are lattice problems in low spatial dimension at finite size, a physical scenario where tensor network methods, both Density Matrix Renormalization Group and beyond, have long proven to be winning strategies. Here we explore in detail the numerical frameworks and methods employed to deal with low-dimension physical setups, from a computational physics perspective. We focus on symmetries and closed-system simulations in arbitrary boundary conditions, while discussing the numerical data structures and linear algebra manipulation routines involved, which form the core libraries of any tensor network code. At a higher level, we put the spotlight on loop-free network geometries, discussing their advantages, and presenting in detail algorithms to simulate low-energy equilibrium states. Accompanied by discussions of data structures, numerical techniques and performance, this anthology serves as a programmer's companion, as well as a self-contained introduction and review of the basic and selected advanced concepts in tensor networks, including examples of their applications.Comment: 115 pages, 56 figure

    Flexible Multi-layer Sparse Approximations of Matrices and Applications

    Get PDF
    The computational cost of many signal processing and machine learning techniques is often dominated by the cost of applying certain linear operators to high-dimensional vectors. This paper introduces an algorithm aimed at reducing the complexity of applying linear operators in high dimension by approximately factorizing the corresponding matrix into few sparse factors. The approach relies on recent advances in non-convex optimization. It is first explained and analyzed in details and then demonstrated experimentally on various problems including dictionary learning for image denoising, and the approximation of large matrices arising in inverse problems

    Randomized methods for matrix computations

    Full text link
    The purpose of this text is to provide an accessible introduction to a set of recently developed algorithms for factorizing matrices. These new algorithms attain high practical speed by reducing the dimensionality of intermediate computations using randomized projections. The algorithms are particularly powerful for computing low-rank approximations to very large matrices, but they can also be used to accelerate algorithms for computing full factorizations of matrices. A key competitive advantage of the algorithms described is that they require less communication than traditional deterministic methods

    Study of compression techniques for partial differential equation solvers

    Get PDF
    Partial Differential Equations (PDEs) are widely applied in many branches of science, and solving them efficiently, from a computational point of view, is one of the cornerstones of modern computational science. The finite element (FE) method is a popular numerical technique for calculating approximate solutions to PDEs. A not necessarily complex finite element analysis containing substructures can easily gen-erate enormous quantities of elements that hinder and slow down simulations. Therefore, compression methods are required to decrease the amount of computational effort while retaining the significant dynamics of the problem. In this study, it was decided to apply a purely algebraic approach. Various methods will be included and discussed, ranging from research-level techniques to other apparently unrelated fields like image compression, via the discrete Fourier transform (DFT) and the Wavelet transform or the Singular Value Decomposition (SVD)

    Compression of unitary rank--structured matrices to CMV-like shape with an application to polynomial rootfinding

    Get PDF
    This paper is concerned with the reduction of a unitary matrix U to CMV-like shape. A Lanczos--type algorithm is presented which carries out the reduction by computing the block tridiagonal form of the Hermitian part of U, i.e., of the matrix U+U^H. By elaborating on the Lanczos approach we also propose an alternative algorithm using elementary matrices which is numerically stable. If U is rank--structured then the same property holds for its Hermitian part and, therefore, the block tridiagonalization process can be performed using the rank--structured matrix technology with reduced complexity. Our interest in the CMV-like reduction is motivated by the unitary and almost unitary eigenvalue problem. In this respect, finally, we discuss the application of the CMV-like reduction for the design of fast companion eigensolvers based on the customary QR iteration
    corecore