964 research outputs found
Fast algorithms for solving H∞-norm minimization problems
We propose an efficient computational approach to minimize the H ∞-norm of a transfer-function matrix depending affinely on a set of free parameters. The minimization problem, formulated as a semi-infinite convex programming problem, is solved via a relaxation approach over a finite set of frequency values. In this way, a significant speed up is achieved by avoiding the solution of high order LMIs resulting by equivalently formulating the minimization problem as a high dimensional semidefinite programming problem. Numerical results illustrate the superiority of proposed approach over LMIs based techniques in solving zero order H∞-norm approximation problems
Large-scale Binary Quadratic Optimization Using Semidefinite Relaxation and Applications
In computer vision, many problems such as image segmentation, pixel
labelling, and scene parsing can be formulated as binary quadratic programs
(BQPs). For submodular problems, cuts based methods can be employed to
efficiently solve large-scale problems. However, general nonsubmodular problems
are significantly more challenging to solve. Finding a solution when the
problem is of large size to be of practical interest, however, typically
requires relaxation. Two standard relaxation methods are widely used for
solving general BQPs--spectral methods and semidefinite programming (SDP), each
with their own advantages and disadvantages. Spectral relaxation is simple and
easy to implement, but its bound is loose. Semidefinite relaxation has a
tighter bound, but its computational complexity is high, especially for large
scale problems. In this work, we present a new SDP formulation for BQPs, with
two desirable properties. First, it has a similar relaxation bound to
conventional SDP formulations. Second, compared with conventional SDP methods,
the new SDP formulation leads to a significantly more efficient and scalable
dual optimization approach, which has the same degree of complexity as spectral
methods. We then propose two solvers, namely, quasi-Newton and smoothing Newton
methods, for the dual problem. Both of them are significantly more efficiently
than standard interior-point methods. In practice, the smoothing Newton solver
is faster than the quasi-Newton solver for dense or medium-sized problems,
while the quasi-Newton solver is preferable for large sparse/structured
problems. Our experiments on a few computer vision applications including
clustering, image segmentation, co-segmentation and registration show the
potential of our SDP formulation for solving large-scale BQPs.Comment: Fixed some typos. 18 pages. Accepted to IEEE Transactions on Pattern
Analysis and Machine Intelligenc
Certification of Real Inequalities -- Templates and Sums of Squares
We consider the problem of certifying lower bounds for real-valued
multivariate transcendental functions. The functions we are dealing with are
nonlinear and involve semialgebraic operations as well as some transcendental
functions like , , , etc. Our general framework is to use
different approximation methods to relax the original problem into polynomial
optimization problems, which we solve by sparse sums of squares relaxations. In
particular, we combine the ideas of the maxplus estimators (originally
introduced in optimal control) and of the linear templates (originally
introduced in static analysis by abstract interpretation). The nonlinear
templates control the complexity of the semialgebraic relaxations at the price
of coarsening the maxplus approximations. In that way, we arrive at a new -
template based - certified global optimization method, which exploits both the
precision of sums of squares relaxations and the scalability of abstraction
methods. We analyze the performance of the method on problems from the global
optimization literature, as well as medium-size inequalities issued from the
Flyspeck project.Comment: 27 pages, 3 figures, 4 table
Certification of Bounds of Non-linear Functions: the Templates Method
The aim of this work is to certify lower bounds for real-valued multivariate
functions, defined by semialgebraic or transcendental expressions. The
certificate must be, eventually, formally provable in a proof system such as
Coq. The application range for such a tool is widespread; for instance Hales'
proof of Kepler's conjecture yields thousands of inequalities. We introduce an
approximation algorithm, which combines ideas of the max-plus basis method (in
optimal control) and of the linear templates method developed by Manna et al.
(in static analysis). This algorithm consists in bounding some of the
constituents of the function by suprema of quadratic forms with a well chosen
curvature. This leads to semialgebraic optimization problems, solved by
sum-of-squares relaxations. Templates limit the blow up of these relaxations at
the price of coarsening the approximation. We illustrate the efficiency of our
framework with various examples from the literature and discuss the interfacing
with Coq.Comment: 16 pages, 3 figures, 2 table
A Coordinate-Descent Algorithm for Tracking Solutions in Time-Varying Optimal Power Flows
Consider a polynomial optimisation problem, whose instances vary continuously
over time. We propose to use a coordinate-descent algorithm for solving such
time-varying optimisation problems. In particular, we focus on relaxations of
transmission-constrained problems in power systems.
On the example of the alternating-current optimal power flows (ACOPF), we
bound the difference between the current approximate optimal cost generated by
our algorithm and the optimal cost for a relaxation using the most recent data
from above by a function of the properties of the instance and the rate of
change to the instance over time. We also bound the number of floating-point
operations that need to be performed between two updates in order to guarantee
the error is bounded from above by a given constant
GMRES-Accelerated ADMM for Quadratic Objectives
We consider the sequence acceleration problem for the alternating direction
method-of-multipliers (ADMM) applied to a class of equality-constrained
problems with strongly convex quadratic objectives, which frequently arise as
the Newton subproblem of interior-point methods. Within this context, the ADMM
update equations are linear, the iterates are confined within a Krylov
subspace, and the General Minimum RESidual (GMRES) algorithm is optimal in its
ability to accelerate convergence. The basic ADMM method solves a
-conditioned problem in iterations. We give
theoretical justification and numerical evidence that the GMRES-accelerated
variant consistently solves the same problem in iterations
for an order-of-magnitude reduction in iterations, despite a worst-case bound
of iterations. The method is shown to be competitive against
standard preconditioned Krylov subspace methods for saddle-point problems. The
method is embedded within SeDuMi, a popular open-source solver for conic
optimization written in MATLAB, and used to solve many large-scale semidefinite
programs with error that decreases like , instead of ,
where is the iteration index.Comment: 31 pages, 7 figures. Accepted for publication in SIAM Journal on
Optimization (SIOPT
- …