31 research outputs found
Approximate level method
In this paper we propose and analyze a variant of the level method [4], which is an algorithm for minimizing nonsmooth convex functions. The main work per iteration is spent on 1) minimizing a piecewise-linear model of the objective function and on 2) projecting onto the intersection of the feasible region and a polyhedron arising as a level set of the model. We show that by replacing exact computations in both cases by approximate computations, in relative scale, the theoretical iteration complexity increases only by the factor of four. This means that while spending less work on the subproblems, we are able to retain the good theoretical properties of the level method.evel method, approximate projections in relative scale, nonsmooth convex optimization, sensitivity analysis, large-scale optimization.
Online and Batch Supervised Background Estimation via L1 Regression
We propose a surprisingly simple model for supervised video background
estimation. Our model is based on regression. As existing methods for
regression do not scale to high-resolution videos, we propose several
simple and scalable methods for solving the problem, including iteratively
reweighted least squares, a homotopy method, and stochastic gradient descent.
We show through extensive experiments that our model and methods match or
outperform the state-of-the-art online and batch methods in virtually all
quantitative and qualitative measures
Parallel Stochastic Newton Method
We propose a parallel stochastic Newton method (PSN) for minimizing
unconstrained smooth convex functions. We analyze the method in the strongly
convex case, and give conditions under which acceleration can be expected when
compared to its serial counterpart. We show how PSN can be applied to the
empirical risk minimization problem, and demonstrate the practical efficiency
of the method through numerical experiments and models of simple matrix
classes
Generalized power method for sparse principal component analysis
In this paper we develop a new approach to sparse principal component analysis (sparse PCA). We propose two single-unit and two block optimization formulations of the sparse PCA problem, aimed at extracting a single sparse dominant principal component of a data matrix, or more components at once, respectively. While the initial formulations involve nonconvex functions, and are therefore computationally intractable, we rewrite them into the form of an optimization program involving maximization of a convex function on a compact set. The dimension of the search space is decreased enormously if the data matrix has many more columns (variables) than rows. We then propose and analyze a simple gradient method suited for the task. It appears that our algorithm has best convergence properties in the case when either the objective function or the feasible set are strongly convex, which is the case with our single-unit formulations and can be enforced in the block case. Finally, we demonstrate numerically on a set of random and gene expression test problems that our approach outperforms existing algorithms both in quality of the obtained solution and in computational speed.sparse PCA, power method, gradient ascent, strongly convex sets, block algorithms.
Stochastic Dual Ascent for Solving Linear Systems
We develop a new randomized iterative algorithm---stochastic dual ascent
(SDA)---for finding the projection of a given vector onto the solution space of
a linear system. The method is dual in nature: with the dual being a
non-strongly concave quadratic maximization problem without constraints. In
each iteration of SDA, a dual variable is updated by a carefully chosen point
in a subspace spanned by the columns of a random matrix drawn independently
from a fixed distribution. The distribution plays the role of a parameter of
the method. Our complexity results hold for a wide family of distributions of
random matrices, which opens the possibility to fine-tune the stochasticity of
the method to particular applications. We prove that primal iterates associated
with the dual process converge to the projection exponentially fast in
expectation, and give a formula and an insightful lower bound for the
convergence rate. We also prove that the same rate applies to dual function
values, primal function values and the duality gap. Unlike traditional
iterative methods, SDA converges under no additional assumptions on the system
(e.g., rank, diagonal dominance) beyond consistency. In fact, our lower bound
improves as the rank of the system matrix drops. Many existing randomized
methods for linear systems arise as special cases of SDA, including randomized
Kaczmarz, randomized Newton, randomized coordinate descent, Gaussian descent,
and their variants. In special cases where our method specializes to a known
algorithm, we either recover the best known rates, or improve upon them.
Finally, we show that the framework can be applied to the distributed average
consensus problem to obtain an array of new algorithms. The randomized gossip
algorithm arises as a special case.Comment: This is a slightly refreshed version of the paper originally
submitted on Dec 21, 2015. We have added a numerical experiment involving
randomized Kaczmarz for rank-deficient systems, added a few relevant
references, and corrected a few typos. Stats: 29 pages, 2 algorithms, 1
figur
Weighted Low-Rank Approximation of Matrices and Background Modeling
We primarily study a special a weighted low-rank approximation of matrices
and then apply it to solve the background modeling problem. We propose two
algorithms for this purpose: one operates in the batch mode on the entire data
and the other one operates in the batch-incremental mode on the data and
naturally captures more background variations and computationally more
effective. Moreover, we propose a robust technique that learns the background
frame indices from the data and does not require any training frames. We
demonstrate through extensive experiments that by inserting a simple weight in
the Frobenius norm, it can be made robust to the outliers similar to the
norm. Our methods match or outperform several state-of-the-art online
and batch background modeling methods in virtually all quantitative and
qualitative measures.Comment: arXiv admin note: text overlap with arXiv:1707.0028
A Batch-Incremental Video Background Estimation Model using Weighted Low-Rank Approximation of Matrices
Principal component pursuit (PCP) is a state-of-the-art approach for
background estimation problems. Due to their higher computational cost, PCP
algorithms, such as robust principal component analysis (RPCA) and its
variants, are not feasible in processing high definition videos. To avoid the
curse of dimensionality in those algorithms, several methods have been proposed
to solve the background estimation problem in an incremental manner. We propose
a batch-incremental background estimation model using a special weighted
low-rank approximation of matrices. Through experiments with real and synthetic
video sequences, we demonstrate that our method is superior to the
state-of-the-art background estimation algorithms such as GRASTA, ReProCS,
incPCP, and GFL