23 research outputs found
An Efficient Approach for Computing Optimal Low-Rank Regularized Inverse Matrices
Standard regularization methods that are used to compute solutions to
ill-posed inverse problems require knowledge of the forward model. In many
real-life applications, the forward model is not known, but training data is
readily available. In this paper, we develop a new framework that uses training
data, as a substitute for knowledge of the forward model, to compute an optimal
low-rank regularized inverse matrix directly, allowing for very fast
computation of a regularized solution. We consider a statistical framework
based on Bayes and empirical Bayes risk minimization to analyze theoretical
properties of the problem. We propose an efficient rank update approach for
computing an optimal low-rank regularized inverse matrix for various error
measures. Numerical experiments demonstrate the benefits and potential
applications of our approach to problems in signal and image processing.Comment: 24 pages, 11 figure
Optimal CUR Matrix Decompositions
The CUR decomposition of an matrix finds an
matrix with a subset of columns of together with an matrix with a subset of rows of as well as a
low-rank matrix such that the matrix approximates the matrix
that is, , where
denotes the Frobenius norm and is the best matrix
of rank constructed via the SVD. We present input-sparsity-time and
deterministic algorithms for constructing such a CUR decomposition where
and and rank. Up to constant
factors, our algorithms are simultaneously optimal in and rank.Comment: small revision in lemma 4.
Generalized brillinger-like transforms
ArtÃculo cientÃficoWe propose novel transforms of stochastic vectors,
called the generalized Brillinger transforms (GBT1 and GBT2),
which are generalizations of the Brillinger transform (BT). The
GBT1 extends the BT to the cases when the covariance matrix
and the weighting matrix are singular, and moreover, the weighting
matrix is not necessarily symmetric. We show that the GBT1
may computationally be preferable over another related optimal
technique, the generic Karhunen–Loève transform (GKLT). The
GBT2 generalizes the GBT1 to provide, under the condition we
impose, better associated accuracy than that of the GBT1. It is
achieved because of the increase in a number of parameters to
optimize compared to that in the GBT1