159 research outputs found

    An Efficient Policy Iteration Algorithm for Dynamic Programming Equations

    Full text link
    We present an accelerated algorithm for the solution of static Hamilton-Jacobi-Bellman equations related to optimal control problems. Our scheme is based on a classic policy iteration procedure, which is known to have superlinear convergence in many relevant cases provided the initial guess is sufficiently close to the solution. In many cases, this limitation degenerates into a behavior similar to a value iteration method, with an increased computation time. The new scheme circumvents this problem by combining the advantages of both algorithms with an efficient coupling. The method starts with a value iteration phase and then switches to a policy iteration procedure when a certain error threshold is reached. A delicate point is to determine this threshold in order to avoid cumbersome computation with the value iteration and, at the same time, to be reasonably sure that the policy iteration method will finally converge to the optimal solution. We analyze the methods and efficient coupling in a number of examples in dimension two, three and four illustrating its properties

    Robust feedback control of nonlinear PDEs by numerical approximation of high-dimensional Hamilton-Jacobi-Isaacs equations

    Get PDF
    Copyright © by SIAM. We propose an approach for the synthesis of robust and optimal feedback controllers for nonlinear PDEs. Our approach considers the approximation of infinite-dimensional control systems by a pseudospectral collocation method, leading to high-dimensional nonlinear dynamics. For the reducedorder model, we construct a robust feedback control based on the H∞ control method, which requires the solution of an associated high-dimensional Hamilton-Jacobi-Isaacs nonlinear PDE. The dimensionality of the Isaacs PDE is tackled by means of a separable representation of the control system, and a polynomial approximation ansatz for the corresponding value function. Our method proves to be effective for the robust stabilization of nonlinear dynamics up to dimension d ≈ 12. We assess the robustness and optimality features of our design over a class of nonlinear parabolic PDEs, including nonlinear advection and reaction terms. The proposed design yields a feedback controller achieving optimal stabilization and disturbance rejection properties, along with providing a modeling framework for the robust control of PDEs under parametric uncertainties

    An efficient policy iteration algorithm for dynamic programming equations

    Get PDF
    We present an accelerated algorithm for the solution of static Hamilton–Jacobi–Bellman equations related to optimal control problems. Our scheme is based on a classic policy iteration procedure, which is known to have superlinear convergence in many relevant cases provided the initial guess is sufficiently close to the solution. This limitation often degenerates into a behavior similar to a value iteration method, with an increased computation time. The new scheme circumvents this problem by combining the advantages of both algorithms with an efficient coupling. The method starts with a coarse-mesh value iteration phase and then switches to a fine-mesh policy iteration procedure when a certain error threshold is reached. A delicate point is to determine this threshold in order to avoid cumbersome computations with the value iteration and at the same time to ensure the convergence of the policy iteration method to the optimal solution. We analyze the methods and efficient coupling in a number of examples in different dimensions, illustrating their properties

    Value iteration convergence of ε-monotone schemes for stationary Hamilton-Jacobi equations

    Get PDF
    International audienceWe present an abstract convergence result for the xed point approximation of stationary Hamilton{Jacobi equations. The basic assumptions on the discrete operator are invariance with respect to the addition of constants, "-monotonicity and consistency. The result can be applied to various high-order approximation schemes which are illustrated in the paper. Several applications to Hamilton{Jacobi equations and numerical tests are presented

    Boundary Treatment and Multigrid Preconditioning for Semi-Lagrangian Schemes Applied to Hamilton-Jacobi-Bellman Equations

    Get PDF
    We analyse two practical aspects that arise in the numerical solution of Hamilton-Jacobi-Bellman (HJB) equations by a particular class of monotone approximation schemes known as semi-Lagrangian schemes. These schemes make use of a wide stencil to achieve convergence and result in discretization matrices that are less sparse and less local than those coming from standard finite difference schemes. This leads to computational difficulties not encountered there. In particular, we consider the overstepping of the domain boundary and analyse the accuracy and stability of stencil truncation. This truncation imposes a stricter CFL condition for explicit schemes in the vicinity of boundaries than in the interior, such that implicit schemes become attractive. We then study the use of geometric, algebraic and aggregation-based multigrid preconditioners to solve the resulting discretised systems from implicit time stepping schemes efficiently. Finally, we illustrate the performance of these techniques numerically for benchmark test cases from the literature

    Deep neural network approximations for the stable manifolds of the Hamilton-Jacobi equations

    Full text link
    As the Riccati equation for control of linear systems, the Hamilton-Jacobi-Bellman (HJB) equations play a fundamental role for optimal control of nonlinear systems. For infinite-horizon optimal control, the stabilizing solution of HJB equation can be represented by the stable manifold of the associated Hamiltonian system. In this paper, we study the neural network (NN) semiglobal approximation of the stable manifold. The main contribution includes two aspects: firstly, from the mathematical point of view, we rigorously prove that if an approximation is sufficiently close to the exact stable manifold of the HJB equation, then the corresponding control derived from this approximation is near optimal. Secondly, we propose a deep learning method to approximate the stable manifolds, and then numerically compute optimal feedback controls. The algorithm is devised from geometric features of the stable manifold, and relies on adaptive data generation by finding trajectories randomly in the stable manifold. The trajectories are found by solving two-point boundary value problems (BVP) locally near the equilibrium and extending the local solution by initial value problems (IVP) for the associated Hamiltonian system. A number of samples are chosen on each trajectory. Some adaptive samples are selected near the points with large errors after the previous round of training. Our algorithm is causality-free basically, hence it has a potential to apply to various high-dimensional nonlinear systems. We illustrate the effectiveness of our method by stabilizing the Reaction Wheel Pendulums.Comment: The algorithm is modified. The main point is that the trajectories on stable manifold are found by a combination of two-point BVP near the equilibrium and initial value problem far away from the equilibrium. The algorithm becomes more effectiv
    • …
    corecore