2,488 research outputs found
Sparse Wide-Area Control of Power Systems using Data-driven Reinforcement Learning
In this paper we present an online wide-area oscillation damping control
(WAC) design for uncertain models of power systems using ideas from
reinforcement learning. We assume that the exact small-signal model of the
power system at the onset of a contingency is not known to the operator and use
the nominal model and online measurements of the generator states and control
inputs to rapidly converge to a state-feedback controller that minimizes a
given quadratic energy cost. However, unlike conventional linear quadratic
regulators (LQR), we intend our controller to be sparse, so its implementation
reduces the communication costs. We, therefore, employ the gradient support
pursuit (GraSP) optimization algorithm to impose sparsity constraints on the
control gain matrix during learning. The sparse controller is thereafter
implemented using distributed communication. Using the IEEE 39-bus power system
model with 1149 unknown parameters, it is demonstrated that the proposed
learning method provides reliable LQR performance while the controller matched
to the nominal model becomes unstable for severely uncertain systems.Comment: Submitted to IEEE ACC 2019. 8 pages, 4 figure
Optimal control of nonlinear partially-unknown systems with unsymmetrical input constraints and its applications to the optimal UAV circumnavigation problem
Aimed at solving the optimal control problem for nonlinear systems with
unsymmetrical input constraints, we present an online adaptive approach for
partially unknown control systems/dynamics. The designed algorithm converges
online to the optimal control solution without the knowledge of the internal
system dynamics. The optimality of the obtained control policy and the
stability for the closed-loop dynamic optimality are proved theoretically. The
proposed method greatly relaxes the assumption on the form of the internal
dynamics and input constraints in previous works. Besides, the control design
framework proposed in this paper offers a new approach to solve the optimal
circumnavigation problem involving a moving target for a fixed-wing unmanned
aerial vehicle (UAV). The control performance of our method is compared with
that of the existing circumnavigation control law in a numerical simulation and
the simulation results validate the effectiveness of our algorithm
Actor-Critic Reinforcement Learning for Control with Stability Guarantee
Reinforcement Learning (RL) and its integration with deep learning have
achieved impressive performance in various robotic control tasks, ranging from
motion planning and navigation to end-to-end visual manipulation. However,
stability is not guaranteed in model-free RL by solely using data. From a
control-theoretic perspective, stability is the most important property for any
control system, since it is closely related to safety, robustness, and
reliability of robotic systems. In this paper, we propose an actor-critic RL
framework for control which can guarantee closed-loop stability by employing
the classic Lyapunov's method in control theory. First of all, a data-based
stability theorem is proposed for stochastic nonlinear systems modeled by
Markov decision process. Then we show that the stability condition could be
exploited as the critic in the actor-critic RL to learn a controller/policy. At
last, the effectiveness of our approach is evaluated on several well-known
3-dimensional robot control tasks and a synthetic biology gene network tracking
task in three different popular physics simulation platforms. As an empirical
evaluation on the advantage of stability, we show that the learned policies can
enable the systems to recover to the equilibrium or way-points when interfered
by uncertainties such as system parametric variations and external disturbances
to a certain extent.Comment: IEEE RA-L + IROS 202
Event-triggered robust control for multi-player nonzero-sum games with input constraints and mismatched uncertainties
In this article, an event-triggered robust control (ETRC) method is investigated for multi-player nonzero-sum games of continuous-time input constrained nonlinear systems with mismatched uncertainties. By constructing an auxiliary system and designing an appropriate value function, the robust control problem of input constrained nonlinear systems is transformed into an optimal regulation problem. Then, a critic neural network (NN) is adopted to approximate the value function of each player for solving the event-triggered coupled Hamilton-Jacobi equation and obtaining control laws. Based on a designed event-triggering condition, control laws are updated when events occur only. Thus, both computational burden and communication bandwidth are reduced. We prove that the weight approximation errors of critic NNs and the closed-loop uncertain multi-player system states are all uniformly ultimately bounded thanks to the Lyapunov's direct method. Finally, two examples are provided to demonstrate the effectiveness of the developed ETRC method
Learning-based Predictive Control for Nonlinear Systems with Unknown Dynamics Subject to Safety Constraints
Model predictive control (MPC) has been widely employed as an effective
method for model-based constrained control. For systems with unknown dynamics,
reinforcement learning (RL) and adaptive dynamic programming (ADP) have
received notable attention to solve the adaptive optimal control problems.
Recently, works on the use of RL in the framework of MPC have emerged, which
can enhance the ability of MPC for data-driven control. However, the safety
under state constraints and the closed-loop robustness are difficult to be
verified due to approximation errors of RL with function approximation
structures. Aiming at the above problem, we propose a data-driven robust MPC
solution based on incremental RL, called data-driven robust learning-based
predictive control (dr-LPC), for perturbed unknown nonlinear systems subject to
safety constraints. A data-driven robust MPC (dr-MPC) is firstly formulated
with a learned predictor. The incremental Dual Heuristic Programming (DHP)
algorithm using an actor-critic architecture is then utilized to solve the
online optimization problem of dr-MPC. In each prediction horizon, the actor
and critic learn time-varying laws for approximating the optimal control policy
and costate respectively, which is different from classical MPCs. The state and
control constraints are enforced in the learning process via building a
Hamilton-Jacobi-Bellman (HJB) equation and a regularized actor-critic learning
structure using logarithmic barrier functions. The closed-loop robustness and
safety of the dr-LPC are proven under function approximation errors. Simulation
results on two control examples have been reported, which show that the dr-LPC
can outperform the DHP and dr-MPC in terms of state regulation, and its average
computational time is much smaller than that with the dr-MPC in both examples.Comment: The paper has been submitted at a IEEE Journal for possible
publicatio
Optimal tracking control for uncertain nonlinear systems with prescribed performance via critic-only ADP
This paper addresses the tracking control problem for a class of nonlinear systems described by Euler-Lagrange equations with uncertain system parameters. The proposed control scheme is capable of guaranteeing prescribed performance from two aspects: 1) A special parameter estimator with prescribed performance properties is embedded in the control scheme. The estimator not only ensures the exponential convergence of the estimation errors under relaxed excitation conditions but also can restrict all estimates to pre-determined bounds during the whole estimation process; 2) The proposed controller can strictly guarantee the user-defined performance specifications on tracking errors, including convergence rate, maximum overshoot, and residual set. More importantly, it has the optimizing ability for the trade-off between performance and control cost. A state transformation method is employed to transform the constrained optimal tracking control problem to an unconstrained stationary optimal problem. Then a critic-only adaptive dynamic programming algorithm is designed to approximate the solution of the Hamilton-Jacobi-Bellman equation and the corresponding optimal control policy. Uniformly ultimately bounded stability is guaranteed via Lyapunov-based stability analysis. Finally, numerical simulation results demonstrate the effectiveness of the proposed control scheme
- …