1,420 research outputs found
A second derivative SQP method: theoretical issues
Sequential quadratic programming (SQP) methods form a class of highly efficient algorithms for solving nonlinearly constrained optimization problems. Although second derivative information may often be calculated, there is little practical theory that justifies exact-Hessian SQP methods. In particular, the resulting quadratic programming (QP) subproblems are often nonconvex, and thus finding their global solutions may be computationally nonviable. This paper presents a second-derivative SQP method based on quadratic subproblems that are either convex, and thus may be solved efficiently, or need not be solved globally. Additionally, an explicit descent-constraint is imposed on certain QP subproblems, which “guides” the iterates through areas in which nonconvexity is a concern. Global convergence of the resulting algorithm is established
Adjoint-based predictor-corrector sequential convex programming for parametric nonlinear optimization
This paper proposes an algorithmic framework for solving parametric
optimization problems which we call adjoint-based predictor-corrector
sequential convex programming. After presenting the algorithm, we prove a
contraction estimate that guarantees the tracking performance of the algorithm.
Two variants of this algorithm are investigated. The first one can be used to
solve nonlinear programming problems while the second variant is aimed to treat
online parametric nonlinear programming problems. The local convergence of
these variants is proved. An application to a large-scale benchmark problem
that originates from nonlinear model predictive control of a hydro power plant
is implemented to examine the performance of the algorithms.Comment: This manuscript consists of 25 pages and 7 figure
A second derivative SQP method: local convergence
In [19], we gave global convergence results for a second-derivative SQP method for minimizing the exact ℓ1-merit function for a fixed value of the penalty parameter. To establish this result, we used the properties of the so-called Cauchy step, which was itself computed from the so-called predictor step. In addition, we allowed for the computation of a variety of (optional) SQP steps that were intended to improve the efficiency of the algorithm. \ud
\ud
Although we established global convergence of the algorithm, we did not discuss certain aspects that are critical when developing software capable of solving general optimization problems. In particular, we must have strategies for updating the penalty parameter and better techniques for defining the positive-definite matrix Bk used in computing the predictor step. In this paper we address both of these issues. We consider two techniques for defining the positive-definite matrix Bk—a simple diagonal approximation and a more sophisticated limited-memory BFGS update. We also analyze a strategy for updating the penalty paramter based on approximately minimizing the ℓ1-penalty function over a sequence of increasing values of the penalty parameter.\ud
\ud
Algorithms based on exact penalty functions have certain desirable properties. To be practical, however, these algorithms must be guaranteed to avoid the so-called Maratos effect. We show that a nonmonotone varient of our algorithm avoids this phenomenon and, therefore, results in asymptotically superlinear local convergence; this is verified by preliminary numerical results on the Hock and Shittkowski test set
A Method to Guarantee Local Convergence for Sequential Quadratic Programming with Poor Hessian Approximation
Sequential Quadratic Programming (SQP) is a powerful class of algorithms for
solving nonlinear optimization problems. Local convergence of SQP algorithms is
guaranteed when the Hessian approximation used in each Quadratic Programming
subproblem is close to the true Hessian. However, a good Hessian approximation
can be expensive to compute. Low cost Hessian approximations only guarantee
local convergence under some assumptions, which are not always satisfied in
practice. To address this problem, this paper proposes a simple method to
guarantee local convergence for SQP with poor Hessian approximation. The
effectiveness of the proposed algorithm is demonstrated in a numerical example
Combining Homotopy Methods and Numerical Optimal Control to Solve Motion Planning Problems
This paper presents a systematic approach for computing local solutions to
motion planning problems in non-convex environments using numerical optimal
control techniques. It extends the range of use of state-of-the-art numerical
optimal control tools to problem classes where these tools have previously not
been applicable. Today these problems are typically solved using motion
planners based on randomized or graph search. The general principle is to
define a homotopy that perturbs, or preferably relaxes, the original problem to
an easily solved problem. By combining a Sequential Quadratic Programming (SQP)
method with a homotopy approach that gradually transforms the problem from a
relaxed one to the original one, practically relevant locally optimal solutions
to the motion planning problem can be computed. The approach is demonstrated in
motion planning problems in challenging 2D and 3D environments, where the
presented method significantly outperforms a state-of-the-art open-source
optimizing sampled-based planner commonly used as benchmark
- …