2,317 research outputs found
On inferring intentions in shared tasks for industrial collaborative robots
Inferring human operators' actions in shared collaborative tasks, plays a crucial role in enhancing the cognitive capabilities of industrial robots. In all these incipient collaborative robotic applications, humans and robots not only should share space but also forces and the execution of a task. In this article, we present a robotic system which is able to identify different human's intentions and to adapt its behavior consequently, only by means of force data. In order to accomplish this aim, three major contributions are presented: (a) force-based operator's intent recognition, (b) force-based dataset of physical human-robot interaction and (c) validation of the whole system in a scenario inspired by a realistic industrial application. This work is an important step towards a more natural and user-friendly manner of physical human-robot interaction in scenarios where humans and robots collaborate in the accomplishment of a task.Peer ReviewedPostprint (published version
Early Turn-taking Prediction with Spiking Neural Networks for Human Robot Collaboration
Turn-taking is essential to the structure of human teamwork. Humans are
typically aware of team members' intention to keep or relinquish their turn
before a turn switch, where the responsibility of working on a shared task is
shifted. Future co-robots are also expected to provide such competence. To that
end, this paper proposes the Cognitive Turn-taking Model (CTTM), which
leverages cognitive models (i.e., Spiking Neural Network) to achieve early
turn-taking prediction. The CTTM framework can process multimodal human
communication cues (both implicit and explicit) and predict human turn-taking
intentions in an early stage. The proposed framework is tested on a simulated
surgical procedure, where a robotic scrub nurse predicts the surgeon's
turn-taking intention. It was found that the proposed CTTM framework
outperforms the state-of-the-art turn-taking prediction algorithms by a large
margin. It also outperforms humans when presented with partial observations of
communication cues (i.e., less than 40% of full actions). This early prediction
capability enables robots to initiate turn-taking actions at an early stage,
which facilitates collaboration and increases overall efficiency.Comment: Submitted to IEEE International Conference on Robotics and Automation
(ICRA) 201
Action Classification in Human Robot Interaction Cells in Manufacturing
Action recognition has become a prerequisite approach to fluent Human-Robot Interaction (HRI) due to a high degree of movement flexibility. With the improvements in machine learning algorithms, robots are gradually transitioning into more human-populated areas. However, HRI systems demand the need for robots to possess enough cognition. The action recognition algorithms require massive training datasets, structural information of objects in the environment, and less expensive models in terms of computational complexity. In addition, many such algorithms are trained on datasets derived from daily activities. The algorithms trained on non-industrial datasets may have an unfavorable impact on implementing models and validating actions in an industrial context. This study proposed a lightweight deep learning model for classifying low-level actions in an assembly setting. The model is based on optical flow feature elicitation and mobilenetV2-SSD action classification and is trained and assessed on an actual industrial activities’ dataset. The experimental outcomes show that the presented method is futuristic and does not require extensive preprocessing; therefore, it can be promising in terms of the feasibility of action recognition for mutual performance monitoring in real-world HRI applications. The test result shows 80% accuracy for low-level RGB action classes. The study’s primary objective is to generate experimental results that may be used as a reference for future HRI algorithms based on the InHard dataset
Comparative performance of human and mobile robotic assistants in collaborative fetch-and-deliver tasks
There is an emerging desire across manufacturing industries to deploy robots that support people in their manual work, rather than replace human workers. This paper explores one such opportunity, which is to field a mobile robotic assistant that travels between part carts and the automotive final assembly line, delivering tools and materials to the human workers. We compare the performance of a mobile robotic assistant to that of a human assistant to gain a better understanding of the factors that impact its effectiveness. Statistically significant differences emerge based on type of assistant, human or robot. Interaction times and idle times are statistically significantly higher for the robotic assistant than the human assistant. We report additional differences in participant's subjective response regarding team fluency, situational awareness, comfort and safety. Finally, we discuss how results from the experiment inform the design of a more effective assistant.BMW Grou
Development of a methodology for the human-robot interaction based on vision systems for collaborative robotics
L'abstract è presente nell'allegato / the abstract is in the attachmen
Application of speed and separation monitoring method in human-robot collaboration: industrial case study
Application of human-robot-collaboration techniques in automotive industries has many advantages on productivity, production quality, and workers’ ergonomy, however workers’ safety aspects play the key role during this collaboration. In this paper, results of the ongoing research about the development of a manufacturing cell for the automotive brake disc assembly that is based on the human-robot collaboration are presented. Operational speed and worker-robot separation monitoring methodology (SSM) as one of the available method to reduce the risk of injury according to the ISO technical specification 15066 on collaborative robot in sharing space with human, has been applied. Virtual environment simulation has been used, considering different percentages of robot maximum speed, to determine the SSM algorithm parameters for estimating the minimum protective distance between the robot and operator. Using ISO/TS 15066 and virtual environment simulation, the minimum separation distance between operator and robot has been estimated. Using human-robot collaboration along with the safety issues specified by SSM system has increased the safety of operation and reduced the operator fatigue during the assembly process
FABRIC: A Framework for the Design and Evaluation of Collaborative Robots with Extended Human Adaptation
A limitation for collaborative robots (cobots) is their lack of ability to
adapt to human partners, who typically exhibit an immense diversity of
behaviors. We present an autonomous framework as a cobot's real-time
decision-making mechanism to anticipate a variety of human characteristics and
behaviors, including human errors, toward a personalized collaboration. Our
framework handles such behaviors in two levels: 1) short-term human behaviors
are adapted through our novel Anticipatory Partially Observable Markov Decision
Process (A-POMDP) models, covering a human's changing intent (motivation),
availability, and capability; 2) long-term changing human characteristics are
adapted by our novel Adaptive Bayesian Policy Selection (ABPS) mechanism that
selects a short-term decision model, e.g., an A-POMDP, according to an estimate
of a human's workplace characteristics, such as her expertise and collaboration
preferences. To design and evaluate our framework over a diversity of human
behaviors, we propose a pipeline where we first train and rigorously test the
framework in simulation over novel human models. Then, we deploy and evaluate
it on our novel physical experiment setup that induces cognitive load on humans
to observe their dynamic behaviors, including their mistakes, and their
changing characteristics such as their expertise. We conduct user studies and
show that our framework effectively collaborates non-stop for hours and adapts
to various changing human behaviors and characteristics in real-time. That
increases the efficiency and naturalness of the collaboration with a higher
perceived collaboration, positive teammate traits, and human trust. We believe
that such an extended human adaptation is key to the long-term use of cobots.Comment: The article is in review for publication in International Journal of
Robotics Researc
- …