2,302 research outputs found
Learning Human-Robot Collaboration Insights through the Integration of Muscle Activity in Interaction Motion Models
Recent progress in human-robot collaboration makes fast and fluid
interactions possible, even when human observations are partial and occluded.
Methods like Interaction Probabilistic Movement Primitives (ProMP) model human
trajectories through motion capture systems. However, such representation does
not properly model tasks where similar motions handle different objects. Under
current approaches, a robot would not adapt its pose and dynamics for proper
handling. We integrate the use of Electromyography (EMG) into the Interaction
ProMP framework and utilize muscular signals to augment the human observation
representation. The contribution of our paper is increased task discernment
when trajectories are similar but tools are different and require the robot to
adjust its pose for proper handling. Interaction ProMPs are used with an
augmented vector that integrates muscle activity. Augmented time-normalized
trajectories are used in training to learn correlation parameters and robot
motions are predicted by finding the best weight combination and temporal
scaling for a task. Collaborative single task scenarios with similar motions
but different objects were used and compared. For one experiment only joint
angles were recorded, for the other EMG signals were additionally integrated.
Task recognition was computed for both tasks. Observation state vectors with
augmented EMG signals were able to completely identify differences across
tasks, while the baseline method failed every time. Integrating EMG signals
into collaborative tasks significantly increases the ability of the system to
recognize nuances in the tasks that are otherwise imperceptible, up to 74.6% in
our studies. Furthermore, the integration of EMG signals for collaboration also
opens the door to a wide class of human-robot physical interactions based on
haptic communication that has been largely unexploited in the field.Comment: 7 pages, 2 figures, 2 tables. As submitted to Humanoids 201
Understanding of Object Manipulation Actions Using Human Multi-Modal Sensory Data
Object manipulation actions represent an important share of the Activities of
Daily Living (ADLs). In this work, we study how to enable service robots to use
human multi-modal data to understand object manipulation actions, and how they
can recognize such actions when humans perform them during human-robot
collaboration tasks. The multi-modal data in this study consists of videos,
hand motion data, applied forces as represented by the pressure patterns on the
hand, and measurements of the bending of the fingers, collected as human
subjects performed manipulation actions. We investigate two different
approaches. In the first one, we show that multi-modal signal (motion, finger
bending and hand pressure) generated by the action can be decomposed into a set
of primitives that can be seen as its building blocks. These primitives are
used to define 24 multi-modal primitive features. The primitive features can in
turn be used as an abstract representation of the multi-modal signal and
employed for action recognition. In the latter approach, the visual features
are extracted from the data using a pre-trained image classification deep
convolutional neural network. The visual features are subsequently used to
train the classifier. We also investigate whether adding data from other
modalities produces a statistically significant improvement in the classifier
performance. We show that both approaches produce a comparable performance.
This implies that image-based methods can successfully recognize human actions
during human-robot collaboration. On the other hand, in order to provide
training data for the robot so it can learn how to perform object manipulation
actions, multi-modal data provides a better alternative
Expressivity in Natural and Artificial Systems
Roboticists are trying to replicate animal behavior in artificial systems.
Yet, quantitative bounds on capacity of a moving platform (natural or
artificial) to express information in the environment are not known. This paper
presents a measure for the capacity of motion complexity -- the expressivity --
of articulated platforms (both natural and artificial) and shows that this
measure is stagnant and unexpectedly limited in extant robotic systems. This
analysis indicates trends in increasing capacity in both internal and external
complexity for natural systems while artificial, robotic systems have increased
significantly in the capacity of computational (internal) states but remained
more or less constant in mechanical (external) state capacity. This work
presents a way to analyze trends in animal behavior and shows that robots are
not capable of the same multi-faceted behavior in rich, dynamic environments as
natural systems.Comment: Rejected from Nature, after review and appeal, July 4, 2018
(submitted May 11, 2018
Feedback Error Learning for Rhythmic Motor Primitives
Abstract — Rhythmic motor primitives can be used to learn a variety of oscillatory behaviors from demonstrations or reward signals, e.g., hopping, walking, running and ball-bouncing. However, frequently, such rhythmic motor primitives lead to failures unless a stabilizing controller ensures their functionality, e.g., a balance controller for a walking gait. As an ideal oscillatory behavior requires the stabilizing controller only for exceptions, e.g., to prevent failures, we devise an online learning approach that reduces the dependence on the stabilizing controller. Inspired by related approaches in model learning, we employ the stabilizing controller’s output as a feedback error learning signal for adapting the gait. We demonstrate the resulting approach in two scenarios: a rhythmic arm’s movements and gait adaptation of an underactuated biped. I
Dynamic Active Constraints for Surgical Robots using Vector Field Inequalities
Robotic assistance allows surgeons to perform dexterous and tremor-free
procedures, but robotic aid is still underrepresented in procedures with
constrained workspaces, such as deep brain neurosurgery and endonasal surgery.
In these procedures, surgeons have restricted vision to areas near the surgical
tooltips, which increases the risk of unexpected collisions between the shafts
of the instruments and their surroundings. In this work, our
vector-field-inequalities method is extended to provide dynamic
active-constraints to any number of robots and moving objects sharing the same
workspace. The method is evaluated with experiments and simulations in which
robot tools have to avoid collisions autonomously and in real-time, in a
constrained endonasal surgical environment. Simulations show that with our
method the combined trajectory error of two robotic systems is optimal.
Experiments using a real robotic system show that the method can autonomously
prevent collisions between the moving robots themselves and between the robots
and the environment. Moreover, the framework is also successfully verified
under teleoperation with tool-tissue interactions.Comment: Accepted on T-RO 2019, 19 Page
In silico case studies of compliant robots: AMARSI deliverable 3.3
In the deliverable 3.2 we presented how the morphological computing ap-
proach can significantly facilitate the control strategy in several scenarios,
e.g. quadruped locomotion, bipedal locomotion and reaching. In particular,
the Kitty experimental platform is an example of the use of morphological
computation to allow quadruped locomotion. In this deliverable we continue
with the simulation studies on the application of the different morphological
computation strategies to control a robotic system
- …