626 research outputs found

    A second derivative SQP method: theoretical issues

    Get PDF
    Sequential quadratic programming (SQP) methods form a class of highly efficient algorithms for solving nonlinearly constrained optimization problems. Although second derivative information may often be calculated, there is little practical theory that justifies exact-Hessian SQP methods. In particular, the resulting quadratic programming (QP) subproblems are often nonconvex, and thus finding their global solutions may be computationally nonviable. This paper presents a second-derivative SQP method based on quadratic subproblems that are either convex, and thus may be solved efficiently, or need not be solved globally. Additionally, an explicit descent-constraint is imposed on certain QP subproblems, which “guides” the iterates through areas in which nonconvexity is a concern. Global convergence of the resulting algorithm is established

    Solving Mathematical Programs with Equilibrium Constraints as Nonlinear Programming: A New Framework

    Full text link
    We present a new framework for the solution of mathematical programs with equilibrium constraints (MPECs). In this algorithmic framework, an MPECs is viewed as a concentration of an unconstrained optimization which minimizes the complementarity measure and a nonlinear programming with general constraints. A strategy generalizing ideas of Byrd-Omojokun's trust region method is used to compute steps. By penalizing the tangential constraints into the objective function, we circumvent the problem of not satisfying MFCQ. A trust-funnel-like strategy is used to balance the improvements on feasibility and optimality. We show that, under MPEC-MFCQ, if the algorithm does not terminate in finite steps, then at least one accumulation point of the iterates sequence is an S-stationary point

    A sequential semidefinite programming method and an application in passive reduced-order modeling

    Full text link
    We consider the solution of nonlinear programs with nonlinear semidefiniteness constraints. The need for an efficient exploitation of the cone of positive semidefinite matrices makes the solution of such nonlinear semidefinite programs more complicated than the solution of standard nonlinear programs. In particular, a suitable symmetrization procedure needs to be chosen for the linearization of the complementarity condition. The choice of the symmetrization procedure can be shifted in a very natural way to certain linear semidefinite subproblems, and can thus be reduced to a well-studied problem. The resulting sequential semidefinite programming (SSP) method is a generalization of the well-known SQP method for standard nonlinear programs. We present a sensitivity result for nonlinear semidefinite programs, and then based on this result, we give a self-contained proof of local quadratic convergence of the SSP method. We also describe a class of nonlinear semidefinite programs that arise in passive reduced-order modeling, and we report results of some numerical experiments with the SSP method applied to problems in that class

    Nonlinear programming without a penalty function or a filter

    Get PDF
    A new method is introduced for solving equality constrained nonlinear optimization problems. This method does not use a penalty function, nor a barrier or a filter, and yet can be proved to be globally convergent to first-order stationary points. It uses different trust-regions to cope with the nonlinearities of the objective function and the constraints, and allows inexact SQP steps that do not lie exactly in the nullspace of the local Jacobian. Preliminary numerical experiments on CUTEr problems indicate that the method performs well

    A new double trust regions SQP method without a penalty function or a filter

    Full text link

    A penalty-function-free line search SQP method for nonlinear programming

    Get PDF
    AbstractWe propose a penalty-function-free non-monotone line search method for nonlinear optimization problems with equality and inequality constraints. This method yields global convergence without using a penalty function or a filter. Each step is required to satisfy a decrease condition for the constraint violation, as well as that for the objective function under some reasonable conditions. The proposed mechanism for accepting steps also combines the non-monotone technique on the decrease condition for the constraint violation, which leads to flexibility and an acceptance behavior comparable with filter based methods. Furthermore, it is shown that the proposed method can avoid the Maratos effect if the search directions are improved by second-order corrections (SOC). So locally superlinear convergence is achieved. We also present some numerical results which confirm the robustness and efficiency of our approach
    • …
    corecore