23,073 research outputs found

    Douglas-Rachford Algorithm for Control- and State-constrained Optimal Control Problems

    Full text link
    We consider the application of the Douglas-Rachford (DR) algorithm to solve linear-quadratic (LQ) control problems with box constraints on the state and control variables. We split the constraints of the optimal control problem into two sets: one involving the ODE with boundary conditions, which is affine, and the other a box. We rewrite the LQ control problems as the minimization of the sum of two convex functions. We find the proximal mappings of these functions which we then employ for the projections in the DR iterations. We propose a numerical algorithm for computing the projection onto the affine set. We present a conjecture for finding the costates and the state constraint multipliers of the optimal control problem, which can in turn be used in verifying the optimality conditions. We carry out numerical experiments with two constrained optimal control problems to illustrate the working and the efficiency of the DR algorithm compared to the traditional approach of direct discretization.Comment: 20 pages, 3 figures, 3 table

    A two-phase gradient method for quadratic programming problems with a single linear constraint and bounds on the variables

    Full text link
    We propose a gradient-based method for quadratic programming problems with a single linear constraint and bounds on the variables. Inspired by the GPCG algorithm for bound-constrained convex quadratic programming [J.J. Mor\'e and G. Toraldo, SIAM J. Optim. 1, 1991], our approach alternates between two phases until convergence: an identification phase, which performs gradient projection iterations until either a candidate active set is identified or no reasonable progress is made, and an unconstrained minimization phase, which reduces the objective function in a suitable space defined by the identification phase, by applying either the conjugate gradient method or a recently proposed spectral gradient method. However, the algorithm differs from GPCG not only because it deals with a more general class of problems, but mainly for the way it stops the minimization phase. This is based on a comparison between a measure of optimality in the reduced space and a measure of bindingness of the variables that are on the bounds, defined by extending the concept of proportioning, which was proposed by some authors for box-constrained problems. If the objective function is bounded, the algorithm converges to a stationary point thanks to a suitable application of the gradient projection method in the identification phase. For strictly convex problems, the algorithm converges to the optimal solution in a finite number of steps even in case of degeneracy. Extensive numerical experiments show the effectiveness of the proposed approach.Comment: 30 pages, 17 figure
    • …
    corecore