31,376 research outputs found
Using fast matrix multiplication to solve structured linear systems
Structured linear algebra techniques are a versatile set of tools;
they enable one to deal at once with various types of matrices, with
features such as Toeplitz-, Hankel-, Vandermonde- or Cauchy-likeness.
Following Kailath, Kung and Morf (1979), the usual way of measuring to
what extent a matrix possesses one such structure is through its
displacement rank, that is, the rank of its image through a suitable
displacement operator. Then, for the families of matrices given above,
the results of Bitmead-Anderson, Morf, Kaltofen, Gohberg-Olshevsky,
Pan (among others) provide algorithm of complexity , up
to logarithmic factors, where is the matrix size and its
displacement rank.
We show that for Toeplitz- Vandermonde-like matrices, this cost can be
reduced to , where is an exponent for
linear algebra. We present consequences for Hermite-Pad\u27e approximation
and bivariate interpolation
An efficient multi-core implementation of a novel HSS-structured multifrontal solver using randomized sampling
We present a sparse linear system solver that is based on a multifrontal
variant of Gaussian elimination, and exploits low-rank approximation of the
resulting dense frontal matrices. We use hierarchically semiseparable (HSS)
matrices, which have low-rank off-diagonal blocks, to approximate the frontal
matrices. For HSS matrix construction, a randomized sampling algorithm is used
together with interpolative decompositions. The combination of the randomized
compression with a fast ULV HSS factorization leads to a solver with lower
computational complexity than the standard multifrontal method for many
applications, resulting in speedups up to 7 fold for problems in our test
suite. The implementation targets many-core systems by using task parallelism
with dynamic runtime scheduling. Numerical experiments show performance
improvements over state-of-the-art sparse direct solvers. The implementation
achieves high performance and good scalability on a range of modern shared
memory parallel systems, including the Intel Xeon Phi (MIC). The code is part
of a software package called STRUMPACK -- STRUctured Matrices PACKage, which
also has a distributed memory component for dense rank-structured matrices
Learning detectors quickly using structured covariance matrices
Computer vision is increasingly becoming interested in the rapid estimation
of object detectors. Canonical hard negative mining strategies are slow as they
require multiple passes of the large negative training set. Recent work has
demonstrated that if the distribution of negative examples is assumed to be
stationary, then Linear Discriminant Analysis (LDA) can learn comparable
detectors without ever revisiting the negative set. Even with this insight,
however, the time to learn a single object detector can still be on the order
of tens of seconds on a modern desktop computer. This paper proposes to
leverage the resulting structured covariance matrix to obtain detectors with
identical performance in orders of magnitude less time and memory. We elucidate
an important connection to the correlation filter literature, demonstrating
that these can also be trained without ever revisiting the negative set
Computational linear algebra over finite fields
We present here algorithms for efficient computation of linear algebra
problems over finite fields
A Fast Algorithm for the Inversion of Quasiseparable Vandermonde-like Matrices
The results on Vandermonde-like matrices were introduced as a generalization
of polynomial Vandermonde matrices, and the displacement structure of these
matrices was used to derive an inversion formula. In this paper we first
present a fast Gaussian elimination algorithm for the polynomial
Vandermonde-like matrices. Later we use the said algorithm to derive fast
inversion algorithms for quasiseparable, semiseparable and well-free
Vandermonde-like matrices having complexity. To do so we
identify structures of displacement operators in terms of generators and the
recurrence relations(2-term and 3-term) between the columns of the basis
transformation matrices for quasiseparable, semiseparable and well-free
polynomials. Finally we present an algorithm to compute the
inversion of quasiseparable Vandermonde-like matrices
Recommended from our members
Preparing sparse solvers for exascale computing.
Sparse solvers provide essential functionality for a wide variety of scientific applications. Highly parallel sparse solvers are essential for continuing advances in high-fidelity, multi-physics and multi-scale simulations, especially as we target exascale platforms. This paper describes the challenges, strategies and progress of the US Department of Energy Exascale Computing project towards providing sparse solvers for exascale computing platforms. We address the demands of systems with thousands of high-performance node devices where exposing concurrency, hiding latency and creating alternative algorithms become essential. The efforts described here are works in progress, highlighting current success and upcoming challenges. This article is part of a discussion meeting issue 'Numerical algorithms for high-performance computational science'
- …