836 research outputs found
Projection methods in conic optimization
There exist efficient algorithms to project a point onto the intersection of
a convex cone and an affine subspace. Those conic projections are in turn the
work-horse of a range of algorithms in conic optimization, having a variety of
applications in science, finance and engineering. This chapter reviews some of
these algorithms, emphasizing the so-called regularization algorithms for
linear conic optimization, and applications in polynomial optimization. This is
a presentation of the material of several recent research articles; we aim here
at clarifying the ideas, presenting them in a general framework, and pointing
out important techniques
Conic Optimization Theory: Convexification Techniques and Numerical Algorithms
Optimization is at the core of control theory and appears in several areas of
this field, such as optimal control, distributed control, system
identification, robust control, state estimation, model predictive control and
dynamic programming. The recent advances in various topics of modern
optimization have also been revamping the area of machine learning. Motivated
by the crucial role of optimization theory in the design, analysis, control and
operation of real-world systems, this tutorial paper offers a detailed overview
of some major advances in this area, namely conic optimization and its emerging
applications. First, we discuss the importance of conic optimization in
different areas. Then, we explain seminal results on the design of hierarchies
of convex relaxations for a wide range of nonconvex problems. Finally, we study
different numerical algorithms for large-scale conic optimization problems.Comment: 18 page
SDPNAL+: A Matlab software for semidefinite programming with bound constraints (version 1.0)
SDPNAL+ is a {\sc Matlab} software package that implements an augmented
Lagrangian based method to solve large scale semidefinite programming problems
with bound constraints. The implementation was initially based on a majorized
semismooth Newton-CG augmented Lagrangian method, here we designed it within an
inexact symmetric Gauss-Seidel based semi-proximal ADMM/ALM (alternating
direction method of multipliers/augmented Lagrangian method) framework for the
purpose of deriving simpler stopping conditions and closing the gap between the
practical implementation of the algorithm and the theoretical algorithm. The
basic code is written in {\sc Matlab}, but some subroutines in C language are
incorporated via Mex files. We also design a convenient interface for users to
input their SDP models into the solver. Numerous problems arising from
combinatorial optimization and binary integer quadratic programming problems
have been tested to evaluate the performance of the solver. Extensive numerical
experiments conducted in [Yang, Sun, and Toh, Mathematical Programming
Computation, 7 (2015), pp. 331--366] show that the proposed method is quite
efficient and robust, in that it is able to solve 98.9\% of the 745 test
instances of SDP problems arising from various applications to the accuracy of
in the relative KKT residual
Templates for Convex Cone Problems with Applications to Sparse Signal Recovery
This paper develops a general framework for solving a variety of convex cone
problems that frequently arise in signal processing, machine learning,
statistics, and other fields. The approach works as follows: first, determine a
conic formulation of the problem; second, determine its dual; third, apply
smoothing; and fourth, solve using an optimal first-order method. A merit of
this approach is its flexibility: for example, all compressed sensing problems
can be solved via this approach. These include models with objective
functionals such as the total-variation norm, ||Wx||_1 where W is arbitrary, or
a combination thereof. In addition, the paper also introduces a number of
technical contributions such as a novel continuation scheme, a novel approach
for controlling the step size, and some new results showing that the smooth and
unsmoothed problems are sometimes formally equivalent. Combined with our
framework, these lead to novel, stable and computationally efficient
algorithms. For instance, our general implementation is competitive with
state-of-the-art methods for solving intensively studied problems such as the
LASSO. Further, numerical experiments show that one can solve the Dantzig
selector problem, for which no efficient large-scale solvers exist, in a few
hundred iterations. Finally, the paper is accompanied with a software release.
This software is not a single, monolithic solver; rather, it is a suite of
programs and routines designed to serve as building blocks for constructing
complete algorithms.Comment: The TFOCS software is available at http://tfocs.stanford.edu This
version has updated reference
- …