17,487 research outputs found
Certification of Real Inequalities -- Templates and Sums of Squares
We consider the problem of certifying lower bounds for real-valued
multivariate transcendental functions. The functions we are dealing with are
nonlinear and involve semialgebraic operations as well as some transcendental
functions like , , , etc. Our general framework is to use
different approximation methods to relax the original problem into polynomial
optimization problems, which we solve by sparse sums of squares relaxations. In
particular, we combine the ideas of the maxplus estimators (originally
introduced in optimal control) and of the linear templates (originally
introduced in static analysis by abstract interpretation). The nonlinear
templates control the complexity of the semialgebraic relaxations at the price
of coarsening the maxplus approximations. In that way, we arrive at a new -
template based - certified global optimization method, which exploits both the
precision of sums of squares relaxations and the scalability of abstraction
methods. We analyze the performance of the method on problems from the global
optimization literature, as well as medium-size inequalities issued from the
Flyspeck project.Comment: 27 pages, 3 figures, 4 table
From error bounds to the complexity of first-order descent methods for convex functions
This paper shows that error bounds can be used as effective tools for
deriving complexity results for first-order descent methods in convex
minimization. In a first stage, this objective led us to revisit the interplay
between error bounds and the Kurdyka-\L ojasiewicz (KL) inequality. One can
show the equivalence between the two concepts for convex functions having a
moderately flat profile near the set of minimizers (as those of functions with
H\"olderian growth). A counterexample shows that the equivalence is no longer
true for extremely flat functions. This fact reveals the relevance of an
approach based on KL inequality. In a second stage, we show how KL inequalities
can in turn be employed to compute new complexity bounds for a wealth of
descent methods for convex problems. Our approach is completely original and
makes use of a one-dimensional worst-case proximal sequence in the spirit of
the famous majorant method of Kantorovich. Our result applies to a very simple
abstract scheme that covers a wide class of descent methods. As a byproduct of
our study, we also provide new results for the globalization of KL inequalities
in the convex framework.
Our main results inaugurate a simple methodology: derive an error bound,
compute the desingularizing function whenever possible, identify essential
constants in the descent method and finally compute the complexity using the
one-dimensional worst case proximal sequence. Our method is illustrated through
projection methods for feasibility problems, and through the famous iterative
shrinkage thresholding algorithm (ISTA), for which we show that the complexity
bound is of the form where the constituents of the bound only depend
on error bound constants obtained for an arbitrary least squares objective with
regularization
- …