2,985 research outputs found
Differential-Algebraic Equations and Beyond: From Smooth to Nonsmooth Constrained Dynamical Systems
The present article presents a summarizing view at differential-algebraic
equations (DAEs) and analyzes how new application fields and corresponding
mathematical models lead to innovations both in theory and in numerical
analysis for this problem class. Recent numerical methods for nonsmooth
dynamical systems subject to unilateral contact and friction illustrate the
topicality of this development.Comment: Preprint of Book Chapte
GMRES-Accelerated ADMM for Quadratic Objectives
We consider the sequence acceleration problem for the alternating direction
method-of-multipliers (ADMM) applied to a class of equality-constrained
problems with strongly convex quadratic objectives, which frequently arise as
the Newton subproblem of interior-point methods. Within this context, the ADMM
update equations are linear, the iterates are confined within a Krylov
subspace, and the General Minimum RESidual (GMRES) algorithm is optimal in its
ability to accelerate convergence. The basic ADMM method solves a
-conditioned problem in iterations. We give
theoretical justification and numerical evidence that the GMRES-accelerated
variant consistently solves the same problem in iterations
for an order-of-magnitude reduction in iterations, despite a worst-case bound
of iterations. The method is shown to be competitive against
standard preconditioned Krylov subspace methods for saddle-point problems. The
method is embedded within SeDuMi, a popular open-source solver for conic
optimization written in MATLAB, and used to solve many large-scale semidefinite
programs with error that decreases like , instead of ,
where is the iteration index.Comment: 31 pages, 7 figures. Accepted for publication in SIAM Journal on
Optimization (SIOPT
An Interior-Point Method with Polynomial Complexity and Superlinear Convergence for Linear Complementarity Problems
For linear programming, a primal-dual interior-point algorithm was recently constructed by Zhang and Tapia that achieves both polynomial complexity andĀ Q-superlinear convergence (Q-quadratic in the nondegenerate case). In this paper, we extend their results to quadratic programming and linear complementarity problems
Interior-point algorithms for convex optimization based on primal-dual metrics
We propose and analyse primal-dual interior-point algorithms for convex
optimization problems in conic form. The families of algorithms we analyse are
so-called short-step algorithms and they match the current best iteration
complexity bounds for primal-dual symmetric interior-point algorithm of
Nesterov and Todd, for symmetric cone programming problems with given
self-scaled barriers. Our results apply to any self-concordant barrier for any
convex cone. We also prove that certain specializations of our algorithms to
hyperbolic cone programming problems (which lie strictly between symmetric cone
programming and general convex optimization problems in terms of generality)
can take advantage of the favourable special structure of hyperbolic barriers.
We make new connections to Riemannian geometry, integrals over operator spaces,
Gaussian quadrature, and strengthen the connection of our algorithms to
quasi-Newton updates and hence first-order methods in general.Comment: 36 page
On affine scaling inexact dogleg methods for bound-constrained nonlinear systems
Within the framework of affine scaling trust-region methods for bound constrained problems, we discuss the use of a inexact dogleg method as a tool for simultaneously handling the trust-region and the bound constraints while seeking for an approximate minimizer of the model. Focusing on bound-constrained systems of nonlinear equations, an inexact affine scaling method for large scale problems, employing the inexact dogleg procedure, is described. Global convergence results are established without any Lipschitz assumption on the Jacobian matrix, and locally fast convergence is shown under standard assumptions. Convergence analysis is performed without specifying the scaling matrix used to handle the bounds, and a rather general class of scaling matrices is allowed in actual algorithms. Numerical results showing the performance of the method are also given
- ā¦