562 research outputs found

    Solving Mathematical Programs with Equilibrium Constraints as Nonlinear Programming: A New Framework

    Full text link
    We present a new framework for the solution of mathematical programs with equilibrium constraints (MPECs). In this algorithmic framework, an MPECs is viewed as a concentration of an unconstrained optimization which minimizes the complementarity measure and a nonlinear programming with general constraints. A strategy generalizing ideas of Byrd-Omojokun's trust region method is used to compute steps. By penalizing the tangential constraints into the objective function, we circumvent the problem of not satisfying MFCQ. A trust-funnel-like strategy is used to balance the improvements on feasibility and optimality. We show that, under MPEC-MFCQ, if the algorithm does not terminate in finite steps, then at least one accumulation point of the iterates sequence is an S-stationary point

    A second derivative SQP method: local convergence

    Get PDF
    In [19], we gave global convergence results for a second-derivative SQP method for minimizing the exact ℓ1-merit function for a fixed value of the penalty parameter. To establish this result, we used the properties of the so-called Cauchy step, which was itself computed from the so-called predictor step. In addition, we allowed for the computation of a variety of (optional) SQP steps that were intended to improve the efficiency of the algorithm. \ud \ud Although we established global convergence of the algorithm, we did not discuss certain aspects that are critical when developing software capable of solving general optimization problems. In particular, we must have strategies for updating the penalty parameter and better techniques for defining the positive-definite matrix Bk used in computing the predictor step. In this paper we address both of these issues. We consider two techniques for defining the positive-definite matrix Bk—a simple diagonal approximation and a more sophisticated limited-memory BFGS update. We also analyze a strategy for updating the penalty paramter based on approximately minimizing the ℓ1-penalty function over a sequence of increasing values of the penalty parameter.\ud \ud Algorithms based on exact penalty functions have certain desirable properties. To be practical, however, these algorithms must be guaranteed to avoid the so-called Maratos effect. We show that a nonmonotone varient of our algorithm avoids this phenomenon and, therefore, results in asymptotically superlinear local convergence; this is verified by preliminary numerical results on the Hock and Shittkowski test set

    Applications of Recurrent Neural Networks to Optimization Problems

    Get PDF

    Optimization and Applications

    Get PDF
    [no abstract available

    A Quasi-Monte-Carlo-Based Feasible Sequential System of Linear Equations Method for Stochastic Programs with Recourse

    Get PDF
    A two-stage stochastic quadratic programming problem with inequality constraints is considered. By quasi-Monte-Carlo-based approximations of the objective function and its first derivative, a feasible sequential system of linear equations method is proposed. A new technique to update the active constraint set is suggested. We show that the sequence generated by the proposed algorithm converges globally to a Karush-Kuhn-Tucker (KKT) point of the problem. In particular, the convergence rate is locally superlinear under some additional conditions
    • …
    corecore