377 research outputs found
Inexact Model: A Framework for Optimization and Variational Inequalities
In this paper we propose a general algorithmic framework for first-order
methods in optimization in a broad sense, including minimization problems,
saddle-point problems and variational inequalities. This framework allows to
obtain many known methods as a special case, the list including accelerated
gradient method, composite optimization methods, level-set methods, proximal
methods. The idea of the framework is based on constructing an inexact model of
the main problem component, i.e. objective function in optimization or operator
in variational inequalities. Besides reproducing known results, our framework
allows to construct new methods, which we illustrate by constructing a
universal method for variational inequalities with composite structure. This
method works for smooth and non-smooth problems with optimal complexity without
a priori knowledge of the problem smoothness. We also generalize our framework
for strongly convex objectives and strongly monotone variational inequalities.Comment: 41 page
A Primal-Dual Algorithmic Framework for Constrained Convex Minimization
We present a primal-dual algorithmic framework to obtain approximate
solutions to a prototypical constrained convex optimization problem, and
rigorously characterize how common structural assumptions affect the numerical
efficiency. Our main analysis technique provides a fresh perspective on
Nesterov's excessive gap technique in a structured fashion and unifies it with
smoothing and primal-dual methods. For instance, through the choices of a dual
smoothing strategy and a center point, our framework subsumes decomposition
algorithms, augmented Lagrangian as well as the alternating direction
method-of-multipliers methods as its special cases, and provides optimal
convergence rates on the primal objective residual as well as the primal
feasibility gap of the iterates for all.Comment: This paper consists of 54 pages with 7 tables and 12 figure
First order algorithms in variational image processing
Variational methods in imaging are nowadays developing towards a quite
universal and flexible tool, allowing for highly successful approaches on tasks
like denoising, deblurring, inpainting, segmentation, super-resolution,
disparity, and optical flow estimation. The overall structure of such
approaches is of the form ; where the functional is a data fidelity term also
depending on some input data and measuring the deviation of from such
and is a regularization functional. Moreover is a (often linear)
forward operator modeling the dependence of data on an underlying image, and
is a positive regularization parameter. While is often
smooth and (strictly) convex, the current practice almost exclusively uses
nonsmooth regularization functionals. The majority of successful techniques is
using nonsmooth and convex functionals like the total variation and
generalizations thereof or -norms of coefficients arising from scalar
products with some frame system. The efficient solution of such variational
problems in imaging demands for appropriate algorithms. Taking into account the
specific structure as a sum of two very different terms to be minimized,
splitting algorithms are a quite canonical choice. Consequently this field has
revived the interest in techniques like operator splittings or augmented
Lagrangians. Here we shall provide an overview of methods currently developed
and recent results as well as some computational studies providing a comparison
of different methods and also illustrating their success in applications.Comment: 60 pages, 33 figure
Variable metric inexact line-search based methods for nonsmooth optimization
We develop a new proximal-gradient method for minimizing the sum of a
differentiable, possibly nonconvex, function plus a convex, possibly non
differentiable, function. The key features of the proposed method are the
definition of a suitable descent direction, based on the proximal operator
associated to the convex part of the objective function, and an Armijo-like
rule to determine the step size along this direction ensuring the sufficient
decrease of the objective function. In this frame, we especially address the
possibility of adopting a metric which may change at each iteration and an
inexact computation of the proximal point defining the descent direction. For
the more general nonconvex case, we prove that all limit points of the iterates
sequence are stationary, while for convex objective functions we prove the
convergence of the whole sequence to a minimizer, under the assumption that a
minimizer exists. In the latter case, assuming also that the gradient of the
smooth part of the objective function is Lipschitz, we also give a convergence
rate estimate, showing the O(1/k) complexity with respect to the function
values. We also discuss verifiable sufficient conditions for the inexact
proximal point and we present the results of a numerical experience on a convex
total variation based image restoration problem, showing that the proposed
approach is competitive with another state-of-the-art method
Decomposition by Partial Linearization: Parallel Optimization of Multi-Agent Systems
We propose a novel decomposition framework for the distributed optimization
of general nonconvex sum-utility functions arising naturally in the system
design of wireless multiuser interfering systems. Our main contributions are:
i) the development of the first class of (inexact) Jacobi best-response
algorithms with provable convergence, where all the users simultaneously and
iteratively solve a suitably convexified version of the original sum-utility
optimization problem; ii) the derivation of a general dynamic pricing mechanism
that provides a unified view of existing pricing schemes that are based,
instead, on heuristics; and iii) a framework that can be easily particularized
to well-known applications, giving rise to very efficient practical (Jacobi or
Gauss-Seidel) algorithms that outperform existing adhoc methods proposed for
very specific problems. Interestingly, our framework contains as special cases
well-known gradient algorithms for nonconvex sum-utility problems, and many
blockcoordinate descent schemes for convex functions.Comment: submitted to IEEE Transactions on Signal Processin
- âŠ