9,894 research outputs found
Conic Descent Redux for Memory-Efficient Optimization
Conic programming has well-documented merits in a gamut of signal processing
and machine learning tasks. This contribution revisits a recently developed
first-order conic descent (CD) solver, and advances it in three aspects:
intuition, theory, and algorithmic implementation. It is found that CD can
afford an intuitive geometric derivation that originates from the dual problem.
This opens the door to novel algorithmic designs, with a momentum variant of
CD, momentum conic descent (MOCO) exemplified. Diving deeper into the dual
behavior CD and MOCO reveals: i) an analytically justified stopping criterion;
and, ii) the potential to design preconditioners to speed up dual convergence.
Lastly, to scale semidefinite programming (SDP) especially for low-rank
solutions, a memory efficient MOCO variant is developed and numerically
validated
Conic Optimization Theory: Convexification Techniques and Numerical Algorithms
Optimization is at the core of control theory and appears in several areas of
this field, such as optimal control, distributed control, system
identification, robust control, state estimation, model predictive control and
dynamic programming. The recent advances in various topics of modern
optimization have also been revamping the area of machine learning. Motivated
by the crucial role of optimization theory in the design, analysis, control and
operation of real-world systems, this tutorial paper offers a detailed overview
of some major advances in this area, namely conic optimization and its emerging
applications. First, we discuss the importance of conic optimization in
different areas. Then, we explain seminal results on the design of hierarchies
of convex relaxations for a wide range of nonconvex problems. Finally, we study
different numerical algorithms for large-scale conic optimization problems.Comment: 18 page
Extended Formulations in Mixed-integer Convex Programming
We present a unifying framework for generating extended formulations for the
polyhedral outer approximations used in algorithms for mixed-integer convex
programming (MICP). Extended formulations lead to fewer iterations of outer
approximation algorithms and generally faster solution times. First, we observe
that all MICP instances from the MINLPLIB2 benchmark library are conic
representable with standard symmetric and nonsymmetric cones. Conic
reformulations are shown to be effective extended formulations themselves
because they encode separability structure. For mixed-integer
conic-representable problems, we provide the first outer approximation
algorithm with finite-time convergence guarantees, opening a path for the use
of conic solvers for continuous relaxations. We then connect the popular
modeling framework of disciplined convex programming (DCP) to the existence of
extended formulations independent of conic representability. We present
evidence that our approach can yield significant gains in practice, with the
solution of a number of open instances from the MINLPLIB2 benchmark library.Comment: To be presented at IPCO 201
Projection methods in conic optimization
There exist efficient algorithms to project a point onto the intersection of
a convex cone and an affine subspace. Those conic projections are in turn the
work-horse of a range of algorithms in conic optimization, having a variety of
applications in science, finance and engineering. This chapter reviews some of
these algorithms, emphasizing the so-called regularization algorithms for
linear conic optimization, and applications in polynomial optimization. This is
a presentation of the material of several recent research articles; we aim here
at clarifying the ideas, presenting them in a general framework, and pointing
out important techniques
- …