24,326 research outputs found
A computational model of human trust in supervisory control of robotic swarms
Trust is an important factor in the interaction between humans and automation to mediate the reliance action of human operators. In this work, we study human factors in supervisory control of robotic swarms and develop a computational model of human trust on swarm systems with varied levels of autonomy (LOA). We extend the classic trust theory by adding an intermediate feedback loop to the trust model, which formulates the human trust evolution as a combination of both open-loop trust anticipation and closed-loop trust feedback. A Kalman filter model is implemented to apply the above structure. We conducted a human experiment to collect user data of supervisory control of robotic swarms. Participants were requested to direct the swarm in a simulated environment to finish a foraging task using control systems with varied LOA. We implement three LOAs: manual, mixed-initiative (MI), and fully autonomous LOA. In the manual and autonomous LOA, swarms are controlled by a human or a search algorithm exclusively, while in the MI LOA, the human operator and algorithm collaboratively control the swarm. We train a personalized model for each participant and evaluate the model performance on a separate data set. Evaluation results show that our Kalman model outperforms existing models including inverse reinforcement learning and dynamic Bayesian network methods.
In summary, the proposed work is novel in the following aspects:
1) This Kalman estimator is the first to model the complete trust evolution process with both closed-loop feedback and open-loop trust anticipation. 2) The proposed model analyzes time-series data to reveal the influence of events that occur during the course of an interaction; namely, a user’s intervention and report of levels of trust. 3) The proposed model considers the operator’s cognitive time lag between perceiving and processing the system display. 4) The proposed model uses the Kalman filter structure to fuse information from different sources to estimate a human operator's mental states. 5) The proposed model provides a personalized model for each individual
Hierarchical Decomposition of Nonlinear Dynamics and Control for System Identification and Policy Distillation
The control of nonlinear dynamical systems remains a major challenge for
autonomous agents. Current trends in reinforcement learning (RL) focus on
complex representations of dynamics and policies, which have yielded impressive
results in solving a variety of hard control tasks. However, this new
sophistication and extremely over-parameterized models have come with the cost
of an overall reduction in our ability to interpret the resulting policies. In
this paper, we take inspiration from the control community and apply the
principles of hybrid switching systems in order to break down complex dynamics
into simpler components. We exploit the rich representational power of
probabilistic graphical models and derive an expectation-maximization (EM)
algorithm for learning a sequence model to capture the temporal structure of
the data and automatically decompose nonlinear dynamics into stochastic
switching linear dynamical systems. Moreover, we show how this framework of
switching models enables extracting hierarchies of Markovian and
auto-regressive locally linear controllers from nonlinear experts in an
imitation learning scenario.Comment: 2nd Annual Conference on Learning for Dynamics and Contro
Data-driven Economic NMPC using Reinforcement Learning
Reinforcement Learning (RL) is a powerful tool to perform data-driven optimal
control without relying on a model of the system. However, RL struggles to
provide hard guarantees on the behavior of the resulting control scheme. In
contrast, Nonlinear Model Predictive Control (NMPC) and Economic NMPC (ENMPC)
are standard tools for the closed-loop optimal control of complex systems with
constraints and limitations, and benefit from a rich theory to assess their
closed-loop behavior. Unfortunately, the performance of (E)NMPC hinges on the
quality of the model underlying the control scheme. In this paper, we show that
an (E)NMPC scheme can be tuned to deliver the optimal policy of the real system
even when using a wrong model. This result also holds for real systems having
stochastic dynamics. This entails that ENMPC can be used as a new type of
function approximator within RL. Furthermore, we investigate our results in the
context of ENMPC and formally connect them to the concept of dissipativity,
which is central for the ENMPC stability. Finally, we detail how these results
can be used to deploy classic RL tools for tuning (E)NMPC schemes. We apply
these tools on both a classical linear MPC setting and a standard nonlinear
example from the ENMPC literature
Sparse Wide-Area Control of Power Systems using Data-driven Reinforcement Learning
In this paper we present an online wide-area oscillation damping control
(WAC) design for uncertain models of power systems using ideas from
reinforcement learning. We assume that the exact small-signal model of the
power system at the onset of a contingency is not known to the operator and use
the nominal model and online measurements of the generator states and control
inputs to rapidly converge to a state-feedback controller that minimizes a
given quadratic energy cost. However, unlike conventional linear quadratic
regulators (LQR), we intend our controller to be sparse, so its implementation
reduces the communication costs. We, therefore, employ the gradient support
pursuit (GraSP) optimization algorithm to impose sparsity constraints on the
control gain matrix during learning. The sparse controller is thereafter
implemented using distributed communication. Using the IEEE 39-bus power system
model with 1149 unknown parameters, it is demonstrated that the proposed
learning method provides reliable LQR performance while the controller matched
to the nominal model becomes unstable for severely uncertain systems.Comment: Submitted to IEEE ACC 2019. 8 pages, 4 figure
On compression rate of quantum autoencoders: Control design, numerical and experimental realization
Quantum autoencoders which aim at compressing quantum information in a
low-dimensional latent space lie in the heart of automatic data compression in
the field of quantum information. In this paper, we establish an upper bound of
the compression rate for a given quantum autoencoder and present a learning
control approach for training the autoencoder to achieve the maximal
compression rate. The upper bound of the compression rate is theoretically
proven using eigen-decomposition and matrix differentiation, which is
determined by the eigenvalues of the density matrix representation of the input
states. Numerical results on 2-qubit and 3-qubit systems are presented to
demonstrate how to train the quantum autoencoder to achieve the theoretically
maximal compression, and the training performance using different machine
learning algorithms is compared. Experimental results of a quantum autoencoder
using quantum optical systems are illustrated for compressing two 2-qubit
states into two 1-qubit states
- …