8,496 research outputs found
Catalyst Acceleration for Gradient-Based Non-Convex Optimization
We introduce a generic scheme to solve nonconvex optimization problems using
gradient-based algorithms originally designed for minimizing convex functions.
Even though these methods may originally require convexity to operate, the
proposed approach allows one to use them on weakly convex objectives, which
covers a large class of non-convex functions typically appearing in machine
learning and signal processing. In general, the scheme is guaranteed to produce
a stationary point with a worst-case efficiency typical of first-order methods,
and when the objective turns out to be convex, it automatically accelerates in
the sense of Nesterov and achieves near-optimal convergence rate in function
values. These properties are achieved without assuming any knowledge about the
convexity of the objective, by automatically adapting to the unknown weak
convexity constant. We conclude the paper by showing promising experimental
results obtained by applying our approach to incremental algorithms such as
SVRG and SAGA for sparse matrix factorization and for learning neural networks
Catalyst Acceleration for First-order Convex Optimization: from Theory to Practice
We introduce a generic scheme for accelerating gradient-based optimization
methods in the sense of Nesterov. The approach, called Catalyst, builds upon
the inexact accelerated proximal point algorithm for minimizing a convex
objective function, and consists of approximately solving a sequence of
well-chosen auxiliary problems, leading to faster convergence. One of the keys
to achieve acceleration in theory and in practice is to solve these
sub-problems with appropriate accuracy by using the right stopping criterion
and the right warm-start strategy. We give practical guidelines to use Catalyst
and present a comprehensive analysis of its global complexity. We show that
Catalyst applies to a large class of algorithms, including gradient descent,
block coordinate descent, incremental algorithms such as SAG, SAGA, SDCA, SVRG,
MISO/Finito, and their proximal variants. For all of these methods, we
establish faster rates using the Catalyst acceleration, for strongly convex and
non-strongly convex objectives. We conclude with extensive experiments showing
that acceleration is useful in practice, especially for ill-conditioned
problems.Comment: link to publisher website:
http://jmlr.org/papers/volume18/17-748/17-748.pd
Accelerated Proximal Stochastic Dual Coordinate Ascent for Regularized Loss Minimization
We introduce a proximal version of the stochastic dual coordinate ascent
method and show how to accelerate the method using an inner-outer iteration
procedure. We analyze the runtime of the framework and obtain rates that
improve state-of-the-art results for various key machine learning optimization
problems including SVM, logistic regression, ridge regression, Lasso, and
multiclass SVM. Experiments validate our theoretical findings
Stochastic forward-backward and primal-dual approximation algorithms with application to online image restoration
Stochastic approximation techniques have been used in various contexts in
data science. We propose a stochastic version of the forward-backward algorithm
for minimizing the sum of two convex functions, one of which is not necessarily
smooth. Our framework can handle stochastic approximations of the gradient of
the smooth function and allows for stochastic errors in the evaluation of the
proximity operator of the nonsmooth function. The almost sure convergence of
the iterates generated by the algorithm to a minimizer is established under
relatively mild assumptions. We also propose a stochastic version of a popular
primal-dual proximal splitting algorithm, establish its convergence, and apply
it to an online image restoration problem.Comment: 5 Figure
Probabilistic models of individual and collective animal behavior
Recent developments in automated tracking allow uninterrupted,
high-resolution recording of animal trajectories, sometimes coupled with the
identification of stereotyped changes of body pose or other behaviors of
interest. Analysis and interpretation of such data represents a challenge: the
timing of animal behaviors may be stochastic and modulated by kinematic
variables, by the interaction with the environment or with the conspecifics
within the animal group, and dependent on internal cognitive or behavioral
state of the individual. Existing models for collective motion typically fail
to incorporate the discrete, stochastic, and internal-state-dependent aspects
of behavior, while models focusing on individual animal behavior typically
ignore the spatial aspects of the problem. Here we propose a probabilistic
modeling framework to address this gap. Each animal can switch stochastically
between different behavioral states, with each state resulting in a possibly
different law of motion through space. Switching rates for behavioral
transitions can depend in a very general way, which we seek to identify from
data, on the effects of the environment as well as the interaction between the
animals. We represent the switching dynamics as a Generalized Linear Model and
show that: (i) forward simulation of multiple interacting animals is possible
using a variant of the Gillespie's Stochastic Simulation Algorithm; (ii)
formulated properly, the maximum likelihood inference of switching rate
functions is tractably solvable by gradient descent; (iii) model selection can
be used to identify factors that modulate behavioral state switching and to
appropriately adjust model complexity to data. To illustrate our framework, we
apply it to two synthetic models of animal motion and to real zebrafish
tracking data.Comment: 26 pages, 11 figure
- …