21 research outputs found

    An Implicit Q Theorem for Hessenberg-like Matrices

    Full text link

    A Parallel Structured Divide-and-Conquer Algorithm for Symmetric Tridiagonal Eigenvalue Problems

    Full text link
    漏 2021 IEEE. Personal use of this material is permitted. Permiss铆on from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertis铆ng or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.[EN] In this article, a parallel structured divide-and-conquer (PSDC) eigensolver is proposed for symmetric tridiagonal matrices based on ScaLAPACK and a parallel structured matrix multiplication algorithm, called PSMMA. Computing the eigenvectors via matrix-matrix multiplications is the most computationally expensive part of the divide-and-conquer algorithm, and one of the matrices involved in such multiplications is a rank-structured Cauchy-like matrix. By exploiting this particular property, PSMMA constructs the local matrices by using generators of Cauchy-like matrices without any communication, and further reduces the computation costs by using a structured low-rank approximation algorithm. Thus, both the communication and computation costs are reduced. Experimental results show that both PSMMA and PSDC are highly scalable and scale to 4096 processes at least. PSDC has better scalability than PHDC that was proposed in [16] and only scaled to 300 processes for the same matrices. Comparing with PDSTEDC in ScaLAPACK, PSDC is always faster and achieves 1.4x-1.6x speedup for some matrices with few deflations. PSDC is also comparable with ELPA, with PSDC being faster than ELPA when using few processes and a little slower when using many processes.The authors would like to thank the referees for their valuable comments which greatly improve the presentation of this article. This work was supported by National Natural Science Foundation of China (No. NNW2019ZT6-B20, NNW2019ZT6B21, NNW2019ZT5-A10, U1611261, 61872392, and U1811461), National Key RD Program of China (2018YFB0204303), NSF of Hunan (No. 2019JJ40339), NSF of NUDT (No. ZK18-03-01), Guangdong Natural Science Foundation (2018B030312002), and the Program for Guangdong Introducing Innovative and Entrepreneurial Teams under Grant 2016ZT06D211. The work of Jose E. Roman was supported by the Spanish Agencia Estatal de Investigacion (AEI) under project SLEPc-DA (PID2019-107379RB-I00).Liao, X.; Li, S.; Lu, Y.; Rom谩n Molt贸, JE. (2021). A Parallel Structured Divide-and-Conquer Algorithm for Symmetric Tridiagonal Eigenvalue Problems. IEEE Transactions on Parallel and Distributed Systems. 32(2):367-378. https://doi.org/10.1109/TPDS.2020.3019471S36737832

    Eigenstructure of order-one-quasiseparable matrices. Three-term and two-term recurrence relations

    Get PDF
    AbstractThis paper presents explicit formulas and algorithms to compute the eigenvalues and eigenvectors of order-one-quasiseparable matrices. Various recursive relations for characteristic polynomials of their principal submatrices are derived. The cost of evaluating the characteristic polynomial of an N脳N matrix and its derivative is only O(N). This leads immediately to several versions of a fast quasiseparable Newton iteration algorithm. In the Hermitian case we extend the Sturm property to the characteristic polynomials of order-one-quasiseparable matrices which yields to several versions of a fast quasiseparable bisection algorithm.Conditions guaranteeing that an eigenvalue of a order-one-quasiseparable matrix is simple are obtained, and an explicit formula for the corresponding eigenvector is derived. The method is further extended to the case when these conditions are not fulfilled. Several particular examples with tridiagonal, (almost) unitary Hessenberg, and Toeplitz matrices are considered.The algorithms are based on new three-term and two-term recurrence relations for the characteristic polynomials of principal submatrices of order-one-quasiseparable matrices R. It turns out that the latter new class of polynomials generalizes and includes two classical families: (i) polynomials orthogonal on the real line (that play a crucial role in a number of classical algorithms in numerical linear algebra), and (ii) the Szeg枚 polynomials (that play a significant role in signal processing). Moreover, new formulas can be seen as generalizations of the classical three-term recurrence relations for the real orthogonal polynomials and of the two-term recurrence relations for the Szeg枚 polynomials

    Factor Fitting, Rank Allocation, and Partitioning in Multilevel Low Rank Matrices

    Full text link
    We consider multilevel low rank (MLR) matrices, defined as a row and column permutation of a sum of matrices, each one a block diagonal refinement of the previous one, with all blocks low rank given in factored form. MLR matrices extend low rank matrices but share many of their properties, such as the total storage required and complexity of matrix-vector multiplication. We address three problems that arise in fitting a given matrix by an MLR matrix in the Frobenius norm. The first problem is factor fitting, where we adjust the factors of the MLR matrix. The second is rank allocation, where we choose the ranks of the blocks in each level, subject to the total rank having a given value, which preserves the total storage needed for the MLR matrix. The final problem is to choose the hierarchical partition of rows and columns, along with the ranks and factors. This paper is accompanied by an open source package that implements the proposed methods

    Structured Eigenvalue Problems

    Get PDF
    Most eigenvalue problems arising in practice are known to be structured. Structure is often introduced by discretization and linearization techniques but may also be a consequence of properties induced by the original problem. Preserving this structure can help preserve physically relevant symmetries in the eigenvalues of the matrix and may improve the accuracy and efficiency of an eigenvalue computation. The purpose of this brief survey is to highlight these facts for some common matrix structures. This includes a treatment of rather general concepts such as structured condition numbers and backward errors as well as an overview of algorithms and applications for several matrix classes including symmetric, skew-symmetric, persymmetric, block cyclic, Hamiltonian, symplectic and orthogonal matrices
    corecore