5,951 research outputs found
Efficient Adjoint Computation for Wavelet and Convolution Operators
First-order optimization algorithms, often preferred for large problems,
require the gradient of the differentiable terms in the objective function.
These gradients often involve linear operators and their adjoints, which must
be applied rapidly. We consider two example problems and derive methods for
quickly evaluating the required adjoint operator. The first example is an image
deblurring problem, where we must compute efficiently the adjoint of
multi-stage wavelet reconstruction. Our formulation of the adjoint works for a
variety of boundary conditions, which allows the formulation to generalize to a
larger class of problems. The second example is a blind channel estimation
problem taken from the optimization literature where we must compute the
adjoint of the convolution of two signals. In each example, we show how the
adjoint operator can be applied efficiently while leveraging existing software.Comment: This manuscript is published in the IEEE Signal Processing Magazine,
Volume 33, Issue 6, November 201
One-Pass Sparsified Gaussian Mixtures
We present a one-pass sparsified Gaussian mixture model (SGMM). Given
data points in dimensions, , the model fits Gaussian distributions
to and (softly) classifies each point to these clusters. After paying an
up-front cost of to precondition the data, we subsample
entries of each data point and discard the full -dimensional data. SGMM
operates in time per iteration for diagonal or spherical
covariances, independent of , while estimating the model parameters in the
full -dimensional space, making it one-pass and hence suitable for streaming
data. We derive the maximum likelihood estimators for the parameters in the
sparsified regime, demonstrate clustering on synthetic and real data, and show
that SGMM is faster than GMM while preserving accuracy.Comment: submitted to IEEE DSW 201
Guarantees for the Kronecker Fast Johnson-Lindenstrauss Transform Using a Coherence and Sampling Argument
In the recent paper [Jin, Kolda & Ward, arXiv:1909.04801], it is proved that
the Kronecker fast Johnson-Lindenstrauss transform (KFJLT) is, in fact, a
Johnson-Lindenstrauss transform, which had previously only been conjectured. In
this paper, we provide an alternative proof of this, for when the KFJLT is
applied to Kronecker vectors, using a coherence and sampling argument. Our
proof yields a different bound on the embedding dimension, which can be
combined with the bound in the paper by Jin et al. to get a better bound
overall. As a stepping stone to proving our result, we also show that the KFJLT
is a subspace embedding for matrices with columns that have Kronecker product
structure. Lastly, we compare the KFJLT to four other sketch techniques in
numerical experiments on both synthetic and real-world data.Comment: Accepted to Linear Algebra and its Application
Preconditioned Data Sparsification for Big Data with Applications to PCA and K-means
We analyze a compression scheme for large data sets that randomly keeps a
small percentage of the components of each data sample. The benefit is that the
output is a sparse matrix and therefore subsequent processing, such as PCA or
K-means, is significantly faster, especially in a distributed-data setting.
Furthermore, the sampling is single-pass and applicable to streaming data. The
sampling mechanism is a variant of previous methods proposed in the literature
combined with a randomized preconditioning to smooth the data. We provide
guarantees for PCA in terms of the covariance matrix, and guarantees for
K-means in terms of the error in the center estimators at a given step. We
present numerical evidence to show both that our bounds are nearly tight and
that our algorithms provide a real benefit when applied to standard test data
sets, as well as providing certain benefits over related sampling approaches.Comment: 28 pages, 10 figure
A quasi-Newton proximal splitting method
A new result in convex analysis on the calculation of proximity operators in
certain scaled norms is derived. We describe efficient implementations of the
proximity calculation for a useful class of functions; the implementations
exploit the piece-wise linear nature of the dual problem. The second part of
the paper applies the previous result to acceleration of convex minimization
problems, and leads to an elegant quasi-Newton method. The optimization method
compares favorably against state-of-the-art alternatives. The algorithm has
extensive applications including signal processing, sparse recovery and machine
learning and classification
Randomized Low-Memory Singular Value Projection
Affine rank minimization algorithms typically rely on calculating the
gradient of a data error followed by a singular value decomposition at every
iteration. Because these two steps are expensive, heuristic approximations are
often used to reduce computational burden. To this end, we propose a recovery
scheme that merges the two steps with randomized approximations, and as a
result, operates on space proportional to the degrees of freedom in the
problem. We theoretically establish the estimation guarantees of the algorithm
as a function of approximation tolerance. While the theoretical approximation
requirements are overly pessimistic, we demonstrate that in practice the
algorithm performs well on the quantum tomography recovery problem.Comment: 13 pages. This version has a revised theorem and new numerical
experiment
- …