10 research outputs found
Efficiently Combining Human Demonstrations and Interventions for Safe Training of Autonomous Systems in Real-Time
This paper investigates how to utilize different forms of human interaction
to safely train autonomous systems in real-time by learning from both human
demonstrations and interventions. We implement two components of the
Cycle-of-Learning for Autonomous Systems, which is our framework for combining
multiple modalities of human interaction. The current effort employs human
demonstrations to teach a desired behavior via imitation learning, then
leverages intervention data to correct for undesired behaviors produced by the
imitation learner to teach novel tasks to an autonomous agent safely, after
only minutes of training. We demonstrate this method in an autonomous perching
task using a quadrotor with continuous roll, pitch, yaw, and throttle commands
and imagery captured from a downward-facing camera in a high-fidelity simulated
environment. Our method improves task completion performance for the same
amount of human interaction when compared to learning from demonstrations
alone, while also requiring on average 32% less data to achieve that
performance. This provides evidence that combining multiple modes of human
interaction can increase both the training speed and overall performance of
policies for autonomous systems.Comment: 9 pages, 6 figure
Gaze-Informed Multi-Objective Imitation Learning from Human Demonstrations
In the field of human-robot interaction, teaching learning agents from human
demonstrations via supervised learning has been widely studied and successfully
applied to multiple domains such as self-driving cars and robot manipulation.
However, the majority of the work on learning from human demonstrations
utilizes only behavioral information from the demonstrator, i.e. what actions
were taken, and ignores other useful information. In particular, eye gaze
information can give valuable insight towards where the demonstrator is
allocating their visual attention, and leveraging such information has the
potential to improve agent performance. Previous approaches have only studied
the utilization of attention in simple, synchronous environments, limiting
their applicability to real-world domains. This work proposes a novel imitation
learning architecture to learn concurrently from human action demonstration and
eye tracking data to solve tasks where human gaze information provides
important context. The proposed method is applied to a visual navigation task,
in which an unmanned quadrotor is trained to search for and navigate to a
target vehicle in a real-world, photorealistic simulated environment. When
compared to a baseline imitation learning architecture, results show that the
proposed gaze augmented imitation learning model is able to learn policies that
achieve significantly higher task completion rates, with more efficient paths,
while simultaneously learning to predict human visual attention. This research
aims to highlight the importance of multimodal learning of visual attention
information from additional human input modalities and encourages the community
to adopt them when training agents from human demonstrations to perform
visuomotor tasks