569 research outputs found
Playing with Duality: An Overview of Recent Primal-Dual Approaches for Solving Large-Scale Optimization Problems
Optimization methods are at the core of many problems in signal/image
processing, computer vision, and machine learning. For a long time, it has been
recognized that looking at the dual of an optimization problem may drastically
simplify its solution. Deriving efficient strategies which jointly brings into
play the primal and the dual problems is however a more recent idea which has
generated many important new contributions in the last years. These novel
developments are grounded on recent advances in convex analysis, discrete
optimization, parallel processing, and non-smooth optimization with emphasis on
sparsity issues. In this paper, we aim at presenting the principles of
primal-dual approaches, while giving an overview of numerical methods which
have been proposed in different contexts. We show the benefits which can be
drawn from primal-dual algorithms both for solving large-scale convex
optimization problems and discrete ones, and we provide various application
examples to illustrate their usefulness
A Class of Randomized Primal-Dual Algorithms for Distributed Optimization
Based on a preconditioned version of the randomized block-coordinate
forward-backward algorithm recently proposed in [Combettes,Pesquet,2014],
several variants of block-coordinate primal-dual algorithms are designed in
order to solve a wide array of monotone inclusion problems. These methods rely
on a sweep of blocks of variables which are activated at each iteration
according to a random rule, and they allow stochastic errors in the evaluation
of the involved operators. Then, this framework is employed to derive
block-coordinate primal-dual proximal algorithms for solving composite convex
variational problems. The resulting algorithm implementations may be useful for
reducing computational complexity and memory requirements. Furthermore, we show
that the proposed approach can be used to develop novel asynchronous
distributed primal-dual algorithms in a multi-agent context
A Stochastic Majorize-Minimize Subspace Algorithm for Online Penalized Least Squares Estimation
Stochastic approximation techniques play an important role in solving many
problems encountered in machine learning or adaptive signal processing. In
these contexts, the statistics of the data are often unknown a priori or their
direct computation is too intensive, and they have thus to be estimated online
from the observed signals. For batch optimization of an objective function
being the sum of a data fidelity term and a penalization (e.g. a sparsity
promoting function), Majorize-Minimize (MM) methods have recently attracted
much interest since they are fast, highly flexible, and effective in ensuring
convergence. The goal of this paper is to show how these methods can be
successfully extended to the case when the data fidelity term corresponds to a
least squares criterion and the cost function is replaced by a sequence of
stochastic approximations of it. In this context, we propose an online version
of an MM subspace algorithm and we study its convergence by using suitable
probabilistic tools. Simulation results illustrate the good practical
performance of the proposed algorithm associated with a memory gradient
subspace, when applied to both non-adaptive and adaptive filter identification
problems
A Non-Local Structure Tensor Based Approach for Multicomponent Image Recovery Problems
Non-Local Total Variation (NLTV) has emerged as a useful tool in variational
methods for image recovery problems. In this paper, we extend the NLTV-based
regularization to multicomponent images by taking advantage of the Structure
Tensor (ST) resulting from the gradient of a multicomponent image. The proposed
approach allows us to penalize the non-local variations, jointly for the
different components, through various matrix norms with .
To facilitate the choice of the hyper-parameters, we adopt a constrained convex
optimization approach in which we minimize the data fidelity term subject to a
constraint involving the ST-NLTV regularization. The resulting convex
optimization problem is solved with a novel epigraphical projection method.
This formulation can be efficiently implemented thanks to the flexibility
offered by recent primal-dual proximal algorithms. Experiments are carried out
for multispectral and hyperspectral images. The results demonstrate the
interest of introducing a non-local structure tensor regularization and show
that the proposed approach leads to significant improvements in terms of
convergence speed over current state-of-the-art methods
Stochastic Approximations and Perturbations in Forward-Backward Splitting for Monotone Operators
We investigate the asymptotic behavior of a stochastic version of the
forward-backward splitting algorithm for finding a zero of the sum of a
maximally monotone set-valued operator and a cocoercive operator in Hilbert
spaces. Our general setting features stochastic approximations of the
cocoercive operator and stochastic perturbations in the evaluation of the
resolvents of the set-valued operator. In addition, relaxations and not
necessarily vanishing proximal parameters are allowed. Weak and strong almost
sure convergence properties of the iterates is established under mild
conditions on the underlying stochastic processes. Leveraging these results, we
also establish the almost sure convergence of the iterates of a stochastic
variant of a primal-dual proximal splitting method for composite minimization
problems
Stochastic Quasi-Fej\'er Block-Coordinate Fixed Point Iterations with Random Sweeping
This work proposes block-coordinate fixed point algorithms with applications
to nonlinear analysis and optimization in Hilbert spaces. The asymptotic
analysis relies on a notion of stochastic quasi-Fej\'er monotonicity, which is
thoroughly investigated. The iterative methods under consideration feature
random sweeping rules to select arbitrarily the blocks of variables that are
activated over the course of the iterations and they allow for stochastic
errors in the evaluation of the operators. Algorithms using quasinonexpansive
operators or compositions of averaged nonexpansive operators are constructed,
and weak and strong convergence results are established for the sequences they
generate. As a by-product, novel block-coordinate operator splitting methods
are obtained for solving structured monotone inclusion and convex minimization
problems. In particular, the proposed framework leads to random
block-coordinate versions of the Douglas-Rachford and forward-backward
algorithms and of some of their variants. In the standard case of block,
our results remain new as they incorporate stochastic perturbations
Stochastic forward-backward and primal-dual approximation algorithms with application to online image restoration
Stochastic approximation techniques have been used in various contexts in
data science. We propose a stochastic version of the forward-backward algorithm
for minimizing the sum of two convex functions, one of which is not necessarily
smooth. Our framework can handle stochastic approximations of the gradient of
the smooth function and allows for stochastic errors in the evaluation of the
proximity operator of the nonsmooth function. The almost sure convergence of
the iterates generated by the algorithm to a minimizer is established under
relatively mild assumptions. We also propose a stochastic version of a popular
primal-dual proximal splitting algorithm, establish its convergence, and apply
it to an online image restoration problem.Comment: 5 Figure
- …