179 research outputs found
Dynamical systems and forward-backward algorithms associated with the sum of a convex subdifferential and a monotone cocoercive operator
In a Hilbert framework, we introduce continuous and discrete dynamical
systems which aim at solving inclusions governed by structured monotone
operators , where is the subdifferential of a
convex lower semicontinuous function , and is a monotone cocoercive
operator. We first consider the extension to this setting of the regularized
Newton dynamic with two potentials. Then, we revisit some related dynamical
systems, namely the semigroup of contractions generated by , and the
continuous gradient projection dynamic. By a Lyapunov analysis, we show the
convergence properties of the orbits of these systems.
The time discretization of these dynamics gives various forward-backward
splitting methods (some new) for solving structured monotone inclusions
involving non-potential terms. The convergence of these algorithms is obtained
under classical step size limitation. Perspectives are given in the field of
numerical splitting methods for optimization, and multi-criteria decision
processes.Comment: 25 page
A Continuous-Time Perspective on Optimal Methods for Monotone Equation Problems
We study \textit{rescaled gradient dynamical systems} in a Hilbert space
, where implicit discretization in a finite-dimensional Euclidean
space leads to high-order methods for solving monotone equations (MEs). Our
framework can be interpreted as a natural generalization of celebrated dual
extrapolation method~\citep{Nesterov-2007-Dual} from first order to high order
via appeal to the regularization toolbox of optimization
theory~\citep{Nesterov-2021-Implementable, Nesterov-2021-Inexact}. More
specifically, we establish the existence and uniqueness of a global solution
and analyze the convergence properties of solution trajectories. We also
present discrete-time counterparts of our high-order continuous-time methods,
and we show that the -order method achieves an ergodic rate of
in terms of a restricted merit function and a pointwise rate
of in terms of a residue function. Under regularity conditions,
the restarted version of -order methods achieves local convergence with
the order . Notably, our methods are \textit{optimal} since they have
matched the lower bound established for solving the monotone equation problems
under a standard linear span assumption~\citep{Lin-2022-Perseus}.Comment: 35 Pages; Add the reference with lower bound constructio
Monotone Inclusions, Acceleration and Closed-Loop Control
We propose and analyze a new dynamical system with a closed-loop control law
in a Hilbert space , aiming to shed light on the acceleration
phenomenon for \textit{monotone inclusion} problems, which unifies a broad
class of optimization, saddle point and variational inequality (VI) problems
under a single framework. Given
that is maximal monotone, we propose a closed-loop control system that is
governed by the operator , where a feedback law
is tuned by the resolution of the algebraic equation
for some
. Our first contribution is to prove the existence and uniqueness
of a global solution via the Cauchy-Lipschitz theorem. We present a simple
Lyapunov function for establishing the weak convergence of trajectories via the
Opial lemma and strong convergence results under additional conditions. We then
prove a global ergodic convergence rate of in terms of a gap
function and a global pointwise convergence rate of in terms of a
residue function. Local linear convergence is established in terms of a
distance function under an error bound condition. Further, we provide an
algorithmic framework based on the implicit discretization of our system in a
Euclidean setting, generalizing the large-step HPE framework. Although the
discrete-time analysis is a simplification and generalization of existing
analyses for a bounded domain, it is largely motivated by the above
continuous-time analysis, illustrating the fundamental role that the
closed-loop control plays in acceleration in monotone inclusion. A highlight of
our analysis is a new result concerning -order tensor algorithms for
monotone inclusion problems, complementing the recent analysis for saddle point
and VI problems.Comment: Accepted by Mathematics of Operations Research; 42 Page
Efficient and Flexible First-Order Optimization Algorithms
Optimization problems occur in many areas in science and engineering. When the optimization problem at hand is of large-scale, the computational cost of the optimization algorithm is a main concern. First-order optimization algorithms—in which updates are performed using only gradient or subgradient of the objective function—have low per-iteration computational cost, which make them suitable for tackling large-scale optimization problems. Even though the per-iteration computational cost of these methods is reasonably low, the number of iterations needed for finding a solution—especially if medium or high accuracy is needed—can in practice be very high; as a result, the overall computational cost of using these methods would still be high. This thesis focuses on one of the most widely used first-order optimization algorithms, namely, the forward–backward splitting algorithm, and attempts to improve its performance. To that end, this thesis proposes novel first-order optimization algorithms which all are built upon the forward–backward method. An important feature of the proposed methods is their flexibility. Using the flexibility of the proposed algorithms along with the safeguarding notion, this thesis provides a framework through which many new and efficient optimization algorithms can be developed. To improve efficiency of the forward–backward algorithm, two main approaches are taken in this thesis. In the first one, a technique is proposed to adjust the point at which the forward–backward operator is evaluated. This is done through including additive terms—which are called deviations—in the input argument of the forward– backward operator. The deviations then, in order to have a convergent algorithm, have to satisfy a safeguard condition at each iteration. Incorporating deviations provides great flexibility to the algorithm and paves the way for designing new and improved forward–backward-based methods. A few instances of employing this flexibility to derive new algorithms are presented in the thesis.In the second proposed approach, a globally (and potentially slow) convergent algorithm can be combined with a fast and locally convergent one to form an efficient optimization scheme. The role of the globally convergent method is to ensure convergence of the overall scheme. The fast local algorithm’s role is to speed up the convergence; this is done by switching from the globally convergent algorithm to the local one whenever it is safe, i.e., when a safeguard condition is satisfied. This approach, which allows for combining different global and local algorithms within its framework, can result in fast and globally convergent optimization schemes
Splitting methods with variable metric for KL functions
We study the convergence of general abstract descent methods applied to a
lower semicontinuous nonconvex function f that satisfies the
Kurdyka-Lojasiewicz inequality in a Hilbert space. We prove that any precompact
sequence converges to a critical point of f and obtain new convergence rates
both for the values and the iterates. The analysis covers alternating versions
of the forward-backward method with variable metric and relative errors. As an
example, a nonsmooth and nonconvex version of the Levenberg-Marquardt algorithm
is detailled
Low Complexity Regularization of Linear Inverse Problems
Inverse problems and regularization theory is a central theme in contemporary
signal processing, where the goal is to reconstruct an unknown signal from
partial indirect, and possibly noisy, measurements of it. A now standard method
for recovering the unknown signal is to solve a convex optimization problem
that enforces some prior knowledge about its structure. This has proved
efficient in many problems routinely encountered in imaging sciences,
statistics and machine learning. This chapter delivers a review of recent
advances in the field where the regularization prior promotes solutions
conforming to some notion of simplicity/low-complexity. These priors encompass
as popular examples sparsity and group sparsity (to capture the compressibility
of natural signals and images), total variation and analysis sparsity (to
promote piecewise regularity), and low-rank (as natural extension of sparsity
to matrix-valued data). Our aim is to provide a unified treatment of all these
regularizations under a single umbrella, namely the theory of partial
smoothness. This framework is very general and accommodates all low-complexity
regularizers just mentioned, as well as many others. Partial smoothness turns
out to be the canonical way to encode low-dimensional models that can be linear
spaces or more general smooth manifolds. This review is intended to serve as a
one stop shop toward the understanding of the theoretical properties of the
so-regularized solutions. It covers a large spectrum including: (i) recovery
guarantees and stability to noise, both in terms of -stability and
model (manifold) identification; (ii) sensitivity analysis to perturbations of
the parameters involved (in particular the observations), with applications to
unbiased risk estimation ; (iii) convergence properties of the forward-backward
proximal splitting scheme, that is particularly well suited to solve the
corresponding large-scale regularized optimization problem
- …