1,342 research outputs found
Projection methods in conic optimization
There exist efficient algorithms to project a point onto the intersection of
a convex cone and an affine subspace. Those conic projections are in turn the
work-horse of a range of algorithms in conic optimization, having a variety of
applications in science, finance and engineering. This chapter reviews some of
these algorithms, emphasizing the so-called regularization algorithms for
linear conic optimization, and applications in polynomial optimization. This is
a presentation of the material of several recent research articles; we aim here
at clarifying the ideas, presenting them in a general framework, and pointing
out important techniques
A Parametric Multi-Convex Splitting Technique with Application to Real-Time NMPC
A novel splitting scheme to solve parametric multiconvex programs is
presented. It consists of a fixed number of proximal alternating minimisations
and a dual update per time step, which makes it attractive in a real-time
Nonlinear Model Predictive Control (NMPC) framework and for distributed
computing environments. Assuming that the parametric program is semi-algebraic
and that its KKT points are strongly regular, a contraction estimate is derived
and it is proven that the sub-optimality error remains stable if two key
parameters are tuned properly. Efficacy of the method is demonstrated by
solving a bilinear NMPC problem to control a DC motor.Comment: To appear in Proceedings of the 53rd IEEE Conference on Decision and
Control 201
A Parametric Non-Convex Decomposition Algorithm for Real-Time and Distributed NMPC
A novel decomposition scheme to solve parametric non-convex programs as they
arise in Nonlinear Model Predictive Control (NMPC) is presented. It consists of
a fixed number of alternating proximal gradient steps and a dual update per
time step. Hence, the proposed approach is attractive in a real-time
distributed context. Assuming that the Nonlinear Program (NLP) is
semi-algebraic and that its critical points are strongly regular, contraction
of the sequence of primal-dual iterates is proven, implying stability of the
sub-optimality error, under some mild assumptions. Moreover, it is shown that
the performance of the optimality-tracking scheme can be enhanced via a
continuation technique. The efficacy of the proposed decomposition method is
demonstrated by solving a centralised NMPC problem to control a DC motor and a
distributed NMPC program for collaborative tracking of unicycles, both within a
real-time framework. Furthermore, an analysis of the sub-optimality error as a
function of the sampling period is proposed given a fixed computational power.Comment: 16 pages, 9 figure
An Alternating Trust Region Algorithm for Distributed Linearly Constrained Nonlinear Programs, Application to the AC Optimal Power Flow
A novel trust region method for solving linearly constrained nonlinear
programs is presented. The proposed technique is amenable to a distributed
implementation, as its salient ingredient is an alternating projected gradient
sweep in place of the Cauchy point computation. It is proven that the algorithm
yields a sequence that globally converges to a critical point. As a result of
some changes to the standard trust region method, namely a proximal
regularisation of the trust region subproblem, it is shown that the local
convergence rate is linear with an arbitrarily small ratio. Thus, convergence
is locally almost superlinear, under standard regularity assumptions. The
proposed method is successfully applied to compute local solutions to
alternating current optimal power flow problems in transmission and
distribution networks. Moreover, the new mechanism for computing a Cauchy point
compares favourably against the standard projected search as for its activity
detection properties
Conic Optimization Theory: Convexification Techniques and Numerical Algorithms
Optimization is at the core of control theory and appears in several areas of
this field, such as optimal control, distributed control, system
identification, robust control, state estimation, model predictive control and
dynamic programming. The recent advances in various topics of modern
optimization have also been revamping the area of machine learning. Motivated
by the crucial role of optimization theory in the design, analysis, control and
operation of real-world systems, this tutorial paper offers a detailed overview
of some major advances in this area, namely conic optimization and its emerging
applications. First, we discuss the importance of conic optimization in
different areas. Then, we explain seminal results on the design of hierarchies
of convex relaxations for a wide range of nonconvex problems. Finally, we study
different numerical algorithms for large-scale conic optimization problems.Comment: 18 page
Optimization Methods for Inverse Problems
Optimization plays an important role in solving many inverse problems.
Indeed, the task of inversion often either involves or is fully cast as a
solution of an optimization problem. In this light, the mere non-linear,
non-convex, and large-scale nature of many of these inversions gives rise to
some very challenging optimization problems. The inverse problem community has
long been developing various techniques for solving such optimization tasks.
However, other, seemingly disjoint communities, such as that of machine
learning, have developed, almost in parallel, interesting alternative methods
which might have stayed under the radar of the inverse problem community. In
this survey, we aim to change that. In doing so, we first discuss current
state-of-the-art optimization methods widely used in inverse problems. We then
survey recent related advances in addressing similar challenges in problems
faced by the machine learning community, and discuss their potential advantages
for solving inverse problems. By highlighting the similarities among the
optimization challenges faced by the inverse problem and the machine learning
communities, we hope that this survey can serve as a bridge in bringing
together these two communities and encourage cross fertilization of ideas.Comment: 13 page
Forward-backward truncated Newton methods for convex composite optimization
This paper proposes two proximal Newton-CG methods for convex nonsmooth
optimization problems in composite form. The algorithms are based on a a
reformulation of the original nonsmooth problem as the unconstrained
minimization of a continuously differentiable function, namely the
forward-backward envelope (FBE). The first algorithm is based on a standard
line search strategy, whereas the second one combines the global efficiency
estimates of the corresponding first-order methods, while achieving fast
asymptotic convergence rates. Furthermore, they are computationally attractive
since each Newton iteration requires the approximate solution of a linear
system of usually small dimension
Solving dynamic stochastic economic models by mathematical programming decomposition methods.
Discrete-time optimal control problems arise naturally in many economic problems. Despite the rapid growth in computing power and new developments in the literature, many economic problems are still quite challenging to solve. Economists are aware of the limitations of some of these approaches for solving these problems due to memory and computational requirements. However, many of the economic models present some special structure that can be exploited in an efficient manner. This paper introduces a decomposition methodology, based on a mathematical programming framework, to compute the equilibrium path in dynamic models by breaking the problem into a set of smaller independent subproblems. We study the performance of the method solving a set of dynamic stochastic economic models. The numerical results reveal that the proposed methodology is efficient in terms of computing time and accuracyDynamic stochastic economic model; Computation of equilibrium; Mathematical programming; Decomposition techniques;
- …