15,997 research outputs found
Challenges in Collaborative HRI for Remote Robot Teams
Collaboration between human supervisors and remote teams of robots is highly
challenging, particularly in high-stakes, distant, hazardous locations, such as
off-shore energy platforms. In order for these teams of robots to truly be
beneficial, they need to be trusted to operate autonomously, performing tasks
such as inspection and emergency response, thus reducing the number of
personnel placed in harm's way. As remote robots are generally trusted less
than robots in close-proximity, we present a solution to instil trust in the
operator through a `mediator robot' that can exhibit social skills, alongside
sophisticated visualisation techniques. In this position paper, we present
general challenges and then take a closer look at one challenge in particular,
discussing an initial study, which investigates the relationship between the
level of control the supervisor hands over to the mediator robot and how this
affects their trust. We show that the supervisor is more likely to have higher
trust overall if their initial experience involves handing over control of the
emergency situation to the robotic assistant. We discuss this result, here, as
well as other challenges and interaction techniques for human-robot
collaboration.Comment: 9 pages. Peer reviewed position paper accepted in the CHI 2019
Workshop: The Challenges of Working on Social Robots that Collaborate with
People (SIRCHI2019), ACM CHI Conference on Human Factors in Computing
Systems, May 2019, Glasgow, U
Effects of spatial ability on multi-robot control tasks
Working with large teams of robots is a very complex and demanding task for any operator and individual differences in spatial ability could significantly affect that performance. In the present study, we examine data from two earlier experiments to investigate the effects of ability for perspective-taking on performance at an urban search and rescue (USAR) task using a realistic simulation and alternate displays. We evaluated the participants' spatial ability using a standard measure of spatial orientation and examined the divergence of performance in accuracy and speed in locating victims, and perceived workload. Our findings show operators with higher spatial ability experienced less workload and marked victims more precisely. An interaction was found for the experimental image queue display for which participants with low spatial ability improved significantly in their accuracy in marking victims over the traditional streaming video display. Copyright 2011 by Human Factors and Ergonomics Society, Inc. All rights reserved
Towards the Safety of Human-in-the-Loop Robotics: Challenges and Opportunities for Safety Assurance of Robotic Co-Workers
The success of the human-robot co-worker team in a flexible manufacturing
environment where robots learn from demonstration heavily relies on the correct
and safe operation of the robot. How this can be achieved is a challenge that
requires addressing both technical as well as human-centric research questions.
In this paper we discuss the state of the art in safety assurance, existing as
well as emerging standards in this area, and the need for new approaches to
safety assurance in the context of learning machines. We then focus on robotic
learning from demonstration, the challenges these techniques pose to safety
assurance and indicate opportunities to integrate safety considerations into
algorithms "by design". Finally, from a human-centric perspective, we stipulate
that, to achieve high levels of safety and ultimately trust, the robotic
co-worker must meet the innate expectations of the humans it works with. It is
our aim to stimulate a discussion focused on the safety aspects of
human-in-the-loop robotics, and to foster multidisciplinary collaboration to
address the research challenges identified
Human-Robot Trust Integrated Task Allocation and Symbolic Motion planning for Heterogeneous Multi-robot Systems
This paper presents a human-robot trust integrated task allocation and motion
planning framework for multi-robot systems (MRS) in performing a set of tasks
concurrently. A set of task specifications in parallel are conjuncted with MRS
to synthesize a task allocation automaton. Each transition of the task
allocation automaton is associated with the total trust value of human in
corresponding robots. Here, the human-robot trust model is constructed with a
dynamic Bayesian network (DBN) by considering individual robot performance,
safety coefficient, human cognitive workload and overall evaluation of task
allocation. Hence, a task allocation path with maximum encoded human-robot
trust can be searched based on the current trust value of each robot in the
task allocation automaton. Symbolic motion planning (SMP) is implemented for
each robot after they obtain the sequence of actions. The task allocation path
can be intermittently updated with this DBN based trust model. The overall
strategy is demonstrated by a simulation with 5 robots and 3 parallel subtask
automata
Mixed Initiative Systems for Human-Swarm Interaction: Opportunities and Challenges
Human-swarm interaction (HSI) involves a number of human factors impacting
human behaviour throughout the interaction. As the technologies used within HSI
advance, it is more tempting to increase the level of swarm autonomy within the
interaction to reduce the workload on humans. Yet, the prospective negative
effects of high levels of autonomy on human situational awareness can hinder
this process. Flexible autonomy aims at trading-off these effects by changing
the level of autonomy within the interaction when required; with
mixed-initiatives combining human preferences and automation's recommendations
to select an appropriate level of autonomy at a certain point of time. However,
the effective implementation of mixed-initiative systems raises fundamental
questions on how to combine human preferences and automation recommendations,
how to realise the selected level of autonomy, and what the future impacts on
the cognitive states of a human are. We explore open challenges that hamper the
process of developing effective flexible autonomy. We then highlight the
potential benefits of using system modelling techniques in HSI by illustrating
how they provide HSI designers with an opportunity to evaluate different
strategies for assessing the state of the mission and for adapting the level of
autonomy within the interaction to maximise mission success metrics.Comment: Author version, accepted at the 2018 IEEE Annual Systems Modelling
Conference, Canberra, Australi
- …