792 research outputs found
Synergy-based policy improvement with path integrals for anthropomorphic hands
In this work, a synergy-based reinforcement learning
algorithm has been developed to confer autonomous grasping
capabilities to anthropomorphic hands. In the presence of
high degrees of freedom, classical machine learning techniques
require a number of iterations that increases with the size of the
problem, thus convergence of the solution is not ensured. The
use of postural synergies determines dimensionality reduction
of the search space and allows recent learning techniques, such
as Policy Improvement with Path Integrals, to become easily
applicable. A key point is the adoption of a suitable reward
function representing the goal of the task and ensuring onestep
performance evaluation. Force-closure quality of the grasp
in the synergies subspace has been chosen as a cost function
for performance evaluation. The experiments conducted on the
SCHUNK 5-Finger Hand demonstrate the effectiveness of the
algorithm showing skills comparable to human capabilities in
learning new grasps and in performing a wide variety from
power to high precision grasps of very small objects
Learning Grasps in a Synergy-based Framework
In this work, a supervised learning strategy has been applied in conjunction with a control strategy to provide anthropomorphic hand-arm systems with autonomous grasping capabilities. Both learning and control algorithms have been developed in a synergy-basedframework in order to address issues related to high dimension of the configuration space, that typically characterizes robotic hands and arms with humanlike kinematics. An experimental setup has been built to learn hand-arm motion from humans during reaching and grasping tasks. Then, a Neural Network (NN) has been realized to generalize the grasps learned by imitation. Since the NN approximates the relationship between the object characteristics and the grasp configuration of the hand-arm system, a synergy-based control strategy has been applied to overcome planning errors. The reach-to-grasp strategy has been tested on a setup constituted by the KUKA LWR 4+Arm and the SCHUNK 5-Finger Hand
Human to robot hand motion mapping methods: review and classification
In this article, the variety of approaches proposed in literature to address the problem of mapping human to robot hand motions are summarized and discussed. We particularly attempt to organize under macro-categories the great quantity of presented methods, that are often difficult to be seen from a general point of view due to different fields of application, specific use of algorithms, terminology and declared goals of the mappings. Firstly, a brief historical overview is reported, in order to provide a look on the emergence of the human to robot hand mapping problem as a both conceptual and analytical challenge that is still open nowadays. Thereafter, the survey mainly focuses on a classification of modern mapping methods under six categories: direct joint, direct Cartesian, taskoriented, dimensionality reduction based, pose recognition based and hybrid mappings. For each of these categories, the general view that associates the related reported studies is provided, and representative references are highlighted. Finally, a concluding discussion along with the authors’ point of view regarding future desirable trends are reported.This work was supported in part by the European Commission’s Horizon 2020 Framework Programme with the project REMODEL under Grant 870133 and in part by the Spanish Government under Grant PID2020-114819GB-I00.Peer ReviewedPostprint (published version
Leveraging Kernelized Synergies on Shared Subspace for Precision Grasping and Dexterous Manipulation
Manipulation in contrast to grasping is a trajectorial task that needs to use dexterous hands. Improving the dexterity of robot hands, increases the controller complexity and thus requires to use the concept of postural synergies. Inspired from postural synergies, this research proposes a new framework called kernelized synergies that focuses on the re-usability of same subspace for precision grasping and dexterous manipulation. In this work, the computed subspace of postural synergies is parameterized by kernelized movement primitives to preserve its grasping and manipulation characteristics and allows its reuse for new objects. The grasp stability of proposed framework is assessed with the force closure quality index, as a cost function. For performance evaluation, the proposed framework is initially tested on two different simulated robot hand models using the Syngrasp toolbox and experimentally, four complex grasping and manipulation tasks are performed and reported. Results confirm the hand agnostic approach of proposed framework and its generalization to distinct objects irrespective of their dimensions
Motion planning using synergies : application to anthropomorphic dual-arm robots
Motion planning is a traditional field in robotics, but new problems are nevertheless incessantly appearing, due to continuous advances in the robot developments. In order to solve these new problems, as well as to improve the existing solutions to classical problems, new approaches are being proposed. A paradigmatic case is the humanoid robotics, since the advances done in this field require motion planners not only to look efficiently for an optimal solution in the classic way, i.e. optimizing consumed energy or time in the plan execution, but also looking for human-like solutions, i.e. requiring the robot movements to be similar to those of the human beings. This anthropomorphism in the robot motion is desired not only for aesthetical reasons, but it is also needed to allow a better and safer human-robot collaboration: humans can predict more easily anthropomorphic robot motions thus avoiding collisions and enhancing the collaboration with the robot. Nevertheless, obtaining a satisfactory performance of these anthropomorphic robotic systems requires the automatic planning of the movements, which is still an arduous and non-evident task since the complexity of the planning problem increases exponentially with the number of degrees of freedom of the robotic system.
This doctoral thesis tackles the problem of planning the motions of dual-arm anthropomorphic robots (optionally with mobile base). The main objective is twofold: obtaining robot motions both in an efficient and in a human-like fashion at the same time. Trying to mimic the human movements while reducing the complexity of the search space for planning purposes leads to the concept of synergies, which could be conceptually defined as correlations (in the joint configuration space as well as in the joint velocity space) between the degrees of freedom of the system. This work proposes new sampling-based motion-planning procedures that exploit the concept of synergies, both in the configuration and velocity space, coordinating the movements of the arms, the hands and the mobile base of mobile anthropomorphic dual-arm robots.La planificación de movimientos es un campo tradicional de la robótica, sin embargo aparecen incesantemente nuevos problemas debido a los continuos avances en el desarrollo de los robots. Para resolver esos nuevos problemas, así como para mejorar las soluciones existentes a los problemas clásicos, se están proponiendo nuevos enfoques. Un caso paradigmático es la robótica humanoide, ya que los avances realizados en este campo requieren que los algoritmos planificadores de movimientos no sólo encuentren eficientemente una solución óptima en el sentido clásico, es decir, optimizar el consumo de energía o el tiempo de ejecución de la trayectoria; sino que también busquen soluciones con apariencia humana, es decir, que el movimiento del robot sea similar al del ser humano. Este antropomorfismo en el movimiento del robot se busca no sólo por razones estéticas, sino porque también es necesario para permitir una colaboración mejor y más segura entre el robot y el operario: el ser humano puede predecir con mayor facilidad los movimientos del robot si éstos son antropomórficos, evitando así las colisiones y mejorando la colaboración humano robot. Sin embargo, para obtener un desempeño satisfactorio de estos sistemas robóticos antropomórficos se requiere una planificación automática de sus movimientos, lo que sigue siendo una tarea ardua y poco evidente, ya que la complejidad del problema aumenta exponencialmente con el número de grados de libertad del sistema robótico. Esta tesis doctoral aborda el problema de la planificación de movimientos en robots antropomorfos bibrazo (opcionalmente con base móvil). El objetivo aquí es doble: obtener movimientos robóticos de forma eficiente y, a la vez, que tengan apariencia humana. Intentar imitar los movimientos humanos mientras a la vez se reduce la complejidad del espacio de búsqueda conduce al concepto de sinergias, que podrían definirse conceptualmente como correlaciones (tanto en el espacio de configuraciones como en el espacio de velocidades de las articulaciones) entre los distintos grados de libertad del sistema. Este trabajo propone nuevos procedimientos de planificación de movimientos que explotan el concepto de sinergias, tanto en el espacio de configuraciones como en el espacio de velocidades, coordinando así los movimientos de los brazos, las manos y la base móvil de robots móviles, bibrazo y antropomórficos.Postprint (published version
Vision Based Adaptation to Kernelized Synergies for Human Inspired Robotic Manipulation
Humans in contrast to robots are excellent in performing fine manipulation
tasks owing to their remarkable dexterity and sensorimotor organization.
Enabling robots to acquire such capabilities, necessitates a framework that not
only replicates the human behaviour but also integrates the multi-sensory
information for autonomous object interaction. To address such limitations,
this research proposes to augment the previously developed kernelized synergies
framework with visual perception to automatically adapt to the unknown objects.
The kernelized synergies, inspired from humans, retain the same reduced
subspace for object grasping and manipulation. To detect object in the scene, a
simplified perception pipeline is used that leverages the RANSAC algorithm with
Euclidean clustering and SVM for object segmentation and recognition
respectively. Further, the comparative analysis of kernelized synergies with
other state of art approaches is made to confirm their flexibility and
effectiveness on the robotic manipulation tasks. The experiments conducted on
the robot hand confirm the robustness of modified kernelized synergies
framework against the uncertainties related to the perception of environment
Active haptic perception in robots: a review
In the past few years a new scenario for robot-based applications has emerged. Service
and mobile robots have opened new market niches. Also, new frameworks for shop-floor
robot applications have been developed. In all these contexts, robots are requested to
perform tasks within open-ended conditions, possibly dynamically varying. These new
requirements ask also for a change of paradigm in the design of robots: on-line and safe
feedback motion control becomes the core of modern robot systems. Future robots will
learn autonomously, interact safely and possess qualities like self-maintenance. Attaining
these features would have been relatively easy if a complete model of the environment
was available, and if the robot actuators could execute motion commands perfectly
relative to this model. Unfortunately, a complete world model is not available and robots
have to plan and execute the tasks in the presence of environmental uncertainties which
makes sensing an important component of new generation robots. For this reason,
today\u2019s new generation robots are equipped with more and more sensing components,
and consequently they are ready to actively deal with the high complexity of the real
world. Complex sensorimotor tasks such as exploration require coordination between the
motor system and the sensory feedback. For robot control purposes, sensory feedback
should be adequately organized in terms of relevant features and the associated data
representation. In this paper, we propose an overall functional picture linking sensing
to action in closed-loop sensorimotor control of robots for touch (hands, fingers). Basic
qualities of haptic perception in humans inspire the models and categories comprising the
proposed classification. The objective is to provide a reasoned, principled perspective on
the connections between different taxonomies used in the Robotics and human haptic
literature. The specific case of active exploration is chosen to ground interesting use
cases. Two reasons motivate this choice. First, in the literature on haptics, exploration has
been treated only to a limited extent compared to grasping and manipulation. Second,
exploration involves specific robot behaviors that exploit distributed and heterogeneous
sensory data
Recommended from our members
Intuitive Human-Machine Interfaces for Non-Anthropomorphic Robotic Hands
As robots become more prevalent in our everyday lives, both in our workplaces and in our homes, it becomes increasingly likely that people who are not experts in robotics will be asked to interface with robotic devices. It is therefore important to develop robotic controls that are intuitive and easy for novices to use. Robotic hands, in particular, are very useful, but their high dimensionality makes creating intuitive human-machine interfaces for them complex. In this dissertation, we study the control of non-anthropomorphic robotic hands by non-roboticists in two contexts: collaborative manipulation and assistive robotics.
In the field of collaborative manipulation, the human and the robot work side by side as independent agents. Teleoperation allows the human to assist the robot when autonomous grasping is not able to deal sufficiently well with corner cases or cannot operate fast enough. Using the teleoperator’s hand as an input device can provide an intuitive control method, but finding a mapping between a human hand and a non-anthropomorphic robot hand can be difficult, due to the hands’ dissimilar kinematics. In this dissertation, we seek to create a mapping between the human hand and a fully actuated, non-anthropomorphic robot hand that is intuitive enough to enable effective real-time teleoperation, even for novice users.
We propose a low-dimensional and continuous teleoperation subspace which can be used as an intermediary for mapping between different hand pose spaces. We first propose the general concept of the subspace, its properties and the variables needed to map from the human hand to a robot hand. We then propose three ways to populate the teleoperation subspace mapping. Two of our mappings use a dataglove to harvest information about the user's hand. We define the mapping between joint space and teleoperation subspace with an empirical definition, which requires a person to define hand motions in an intuitive, hand-specific way, and with an algorithmic definition, which is kinematically independent, and uses objects to define the subspace. Our third mapping for the teleoperation subspace uses forearm electromyography (EMG) as a control input.
Assistive orthotics is another area of robotics where human-machine interfaces are critical, since, in this field, the robot is attached to the hand of the human user. In this case, the goal is for the robot to assist the human with movements they would not otherwise be able to achieve. Orthotics can improve the quality of life of people who do not have full use of their hands. Human-machine interfaces for assistive hand orthotics that use EMG signals from the affected forearm as input are intuitive and repeated use can strengthen the muscles of the user's affected arm. In this dissertation, we seek to create an EMG based control for an orthotic device used by people who have had a stroke. We would like our control to enable functional motions when used in conjunction with a orthosis and to be robust to changes in the input signal.
We propose a control for a wearable hand orthosis which uses an easy to don, commodity forearm EMG band. We develop an supervised algorithm to detect a user’s intent to open and close their hand, and pair this algorithm with a training protocol which makes our intent detection robust to changes in the input signal. We show that this algorithm, when used in conjunction with an orthosis over several weeks, can improve distal function in users. Additionally, we propose two semi-supervised intent detection algorithms designed to keep our control robust to changes in the input data while reducing the length and frequency of our training protocol
- …