152 research outputs found

    Faster Sparse Matrix Inversion and Rank Computation in Finite Fields

    Full text link
    We improve the current best running time value to invert sparse matrices over finite fields, lowering it to an expected O(n2.2131)O\big(n^{2.2131}\big) time for the current values of fast rectangular matrix multiplication. We achieve the same running time for the computation of the rank and nullspace of a sparse matrix over a finite field. This improvement relies on two key techniques. First, we adopt the decomposition of an arbitrary matrix into block Krylov and Hankel matrices from Eberly et al. (ISSAC 2007). Second, we show how to recover the explicit inverse of a block Hankel matrix using low displacement rank techniques for structured matrices and fast rectangular matrix multiplication algorithms. We generalize our inversion method to block structured matrices with other displacement operators and strengthen the best known upper bounds for explicit inversion of block Toeplitz-like and block Hankel-like matrices, as well as for explicit inversion of block Vandermonde-like matrices with structured blocks. As a further application, we improve the complexity of several algorithms in topological data analysis and in finite group theory

    Faster Sparse Matrix Inversion and Rank Computation in Finite Fields

    Get PDF
    We improve the current best running time value to invert sparse matrices over finite fields, lowering it to an expected O(n^{2.2131}) time for the current values of fast rectangular matrix multiplication. We achieve the same running time for the computation of the rank and nullspace of a sparse matrix over a finite field. This improvement relies on two key techniques. First, we adopt the decomposition of an arbitrary matrix into block Krylov and Hankel matrices from Eberly et al. (ISSAC 2007). Second, we show how to recover the explicit inverse of a block Hankel matrix using low displacement rank techniques for structured matrices and fast rectangular matrix multiplication algorithms. We generalize our inversion method to block structured matrices with other displacement operators and strengthen the best known upper bounds for explicit inversion of block Toeplitz-like and block Hankel-like matrices, as well as for explicit inversion of block Vandermonde-like matrices with structured blocks. As a further application, we improve the complexity of several algorithms in topological data analysis and in finite group theory

    Development of symbolic algorithms for certain algebraic processes

    Get PDF
    This study investigates the problem of computing the exact greatest common divisor of two polynomials relative to an orthogonal basis, defined over the rational number field. The main objective of the study is to design and implement an effective and efficient symbolic algorithm for the general class of dense polynomials, given the rational number defining terms of their basis. From a general algorithm using the comrade matrix approach, the nonmodular and modular techniques are prescribed. If the coefficients of the generalized polynomials are multiprecision integers, multiprecision arithmetic will be required in the construction of the comrade matrix and the corresponding systems coefficient matrix. In addition, the application of the nonmodular elimination technique on this coefficient matrix extensively applies multiprecision rational number operations. The modular technique is employed to minimize the complexity involved in such computations. A divisor test algorithm that enables the detection of an unlucky reduction is a crucial device for an effective implementation of the modular technique. With the bound of the true solution not known a priori, the test is devised and carefully incorporated into the modular algorithm. The results illustrate that the modular algorithm illustrate its best performance for the class of relatively prime polynomials. The empirical computing time results show that the modular algorithm is markedly superior to the nonmodular algorithms in the case of sufficiently dense Legendre basis polynomials with a small GCD solution. In the case of dense Legendre basis polynomials with a big GCD solution, the modular algorithm is significantly superior to the nonmodular algorithms in higher degree polynomials. For more definitive conclusions, the computing time functions of the algorithms that are presented in this report have been worked out. Further investigations have also been suggested

    On the exact evaluation of integrals of wavelets

    Get PDF
    Wavelet expansions are a powerful tool for constructing adaptive approximations. For this reason, they find applications in a variety of fields, from signal processing to approximation theory. Wavelets are usually derived from refinable functions, which are the solution of a recursive functional equation called the refinement equation. The analytical expression of refinable functions is known in only a few cases, so if we need to evaluate refinable functions we can make use only of the refinement equation. This is also true for the evaluation of their derivatives and integrals. In this paper, we detail a procedure for computing integrals of wavelet products exactly, up to machine precision. The efficient and accurate evaluation of these integrals is particularly required for the computation of the connection coefficients in the wavelet Galerkin method. We show the effectiveness of the procedure by evaluating the integrals of pseudo-splines

    Adaptive isogeometric methods with hierarchical splines: Optimality and convergence rates

    Get PDF
    We consider an adaptive isogeometric method (AIGM) based on (truncated) hierarchical B-splines and continue the study of its numerical properties. We prove that our AIGM is optimal in the sense that delivers optimal convergence rates as soon as the solution of the underlying partial differential equation belongs to a suitable approximation class. The main tool we use is the theory of adaptive methods, together with a local upper bound for the residual error indicators based on suitable properties of a well selected quasi-interpolation operator on hierarchical spline spaces

    TR-2012001: Algebraic Algorithms

    Full text link

    TR-2013009: Algebraic Algorithms

    Full text link
    • …
    corecore