2,849 research outputs found

    Hamiltonian-Driven Adaptive Dynamic Programming with Efficient Experience Replay

    Get PDF
    This article presents a novel efficient experience-replay-based adaptive dynamic programming (ADP) for the optimal control problem of a class of nonlinear dynamical systems within the Hamiltonian-driven framework. The quasi-Hamiltonian is presented for the policy evaluation problem with an admissible policy. With the quasi-Hamiltonian, a novel composite critic learning mechanism is developed to combine the instantaneous data with the historical data. In addition, the pseudo-Hamiltonian is defined to deal with the performance optimization problem. Based on the pseudo-Hamiltonian, the conventional Hamilton–Jacobi–Bellman (HJB) equation can be represented in a filtered form, which can be implemented online. Theoretical analysis is investigated in terms of the convergence of the adaptive critic design and the stability of the closed-loop systems, where parameter convergence can be achieved under a weakened excitation condition. Simulation studies are investigated to verify the efficacy of the presented design scheme

    Self-Evaluation Applied Mathematics 2003-2008 University of Twente

    Get PDF
    This report contains the self-study for the research assessment of the Department of Applied Mathematics (AM) of the Faculty of Electrical Engineering, Mathematics and Computer Science (EEMCS) at the University of Twente (UT). The report provides the information for the Research Assessment Committee for Applied Mathematics, dealing with mathematical sciences at the three universities of technology in the Netherlands. It describes the state of affairs pertaining to the period 1 January 2003 to 31 December 2008

    Quantum feedback control and classical control theory

    Get PDF
    We introduce and discuss the problem of quantum feedback control in the context of established formulations of classical control theory, examining conceptual analogies and essential differences. We describe the application of state-observer-based control laws, familiar in classical control theory, to quantum systems and apply our methods to the particular case of switching the state of a particle in a double-well potential

    Data-Efficient Reinforcement Learning with Probabilistic Model Predictive Control

    Full text link
    Trial-and-error based reinforcement learning (RL) has seen rapid advancements in recent times, especially with the advent of deep neural networks. However, the majority of autonomous RL algorithms require a large number of interactions with the environment. A large number of interactions may be impractical in many real-world applications, such as robotics, and many practical systems have to obey limitations in the form of state space or control constraints. To reduce the number of system interactions while simultaneously handling constraints, we propose a model-based RL framework based on probabilistic Model Predictive Control (MPC). In particular, we propose to learn a probabilistic transition model using Gaussian Processes (GPs) to incorporate model uncertainty into long-term predictions, thereby, reducing the impact of model errors. We then use MPC to find a control sequence that minimises the expected long-term cost. We provide theoretical guarantees for first-order optimality in the GP-based transition models with deterministic approximate inference for long-term planning. We demonstrate that our approach does not only achieve state-of-the-art data efficiency, but also is a principled way for RL in constrained environments.Comment: Accepted at AISTATS 2018

    Optimal fluctuations and the control of chaos.

    Get PDF
    The energy-optimal migration of a chaotic oscillator from one attractor to another coexisting attractor is investigated via an analogy between the Hamiltonian theory of fluctuations and Hamiltonian formulation of the control problem. We demonstrate both on physical grounds and rigorously that the Wentzel-Freidlin Hamiltonian arising in the analysis of fluctuations is equivalent to Pontryagin's Hamiltonian in the control problem with an additive linear unrestricted control. The deterministic optimal control function is identied with the optimal fluctuational force. Numerical and analogue experiments undertaken to verify these ideas demonstrate that, in the limit of small noise intensity, fluctuational escape from the chaotic attractor occurs via a unique (optimal) path corresponding to a unique (optimal) fluctuational force. Initial conditions on the chaotic attractor are identified. The solution of the boundary value control problem for the Pontryagin Hamiltonian is found numerically. It is shown that this solution is approximated very accurately by the optimal fluctuational force found using statistical analysis of the escape trajectories. A second series of numerical experiments on the deterministic system (i.e. in the absence of noise) show that a control function of precisely the same shape and magnitude is indeed able to instigate escape. It is demonstrated that this control function minimizes the cost functional and the corresponding energy is found to be smaller than that obtained with some earlier adaptive control algorithms
    • …
    corecore