6 research outputs found
Online inverse reinforcement learning with unknown disturbances
This paper addresses the problem of online inverse reinforcement learning for
nonlinear systems with modeling uncertainties while in the presence of unknown
disturbances. The developed approach observes state and input trajectories for
an agent and identifies the unknown reward function online. Sub-optimality
introduced in the observed trajectories by the unknown external disturbance is
compensated for using a novel model-based inverse reinforcement learning
approach. The observer estimates the external disturbances and uses the
resulting estimates to learn the dynamic model of the demonstrator. The learned
demonstrator model along with the observed suboptimal trajectories are used to
implement inverse reinforcement learning. Theoretical guarantees are provided
using Lyapunov theory and a simulation example is shown to demonstrate the
effectiveness of the proposed technique.Comment: 8 pages, 3 figure
Online Observer-Based Inverse Reinforcement Learning
In this paper, a novel approach to the output-feedback inverse reinforcement
learning (IRL) problem is developed by casting the IRL problem, for linear
systems with quadratic cost functions, as a state estimation problem. Two
observer-based techniques for IRL are developed, including a novel observer
method that re-uses previous state estimates via history stacks. Theoretical
guarantees for convergence and robustness are established under appropriate
excitation conditions. Simulations demonstrate the performance of the developed
observers and filters under noisy and noise-free measurements.Comment: 7 pages, 3 figure
Reinforcement Learning, Intelligent Control and their Applications in Connected and Autonomous Vehicles
Reinforcement learning (RL) has attracted large attention over the past few years. Recently, we developed a data-driven algorithm to solve predictive cruise control (PCC) and games output regulation problems. This work integrates our recent contributions to the application of RL in game theory, output regulation problems, robust control, small-gain theory and PCC. The algorithm was developed for adaptive optimal output regulation of uncertain linear systems, and uncertain partially linear systems to reject disturbance and also force the output of the systems to asymptotically track a reference. In the PCC problem, we determined the reference velocity for each autonomous vehicle in the platoon using the traffic information broadcasted from the lights to reduce the vehicles\u27 trip time. Then we employed the algorithm to design an approximate optimal controller for the vehicles. This controller is able to regulate the headway, velocity and acceleration of each vehicle to the desired values. Simulation results validate the effectiveness of the algorithms