8,449 research outputs found
On the validation of SPDM task verification facility
This paper describes a methodology for validating a ground-based, hardware-in-the-loop, space-robot simulation facility. This facility, called ‘‘SPDM task verification facility,’’ is being developed by the Canadian Space Agency for the purpose of verifying the contact dynamics performance of the special purpose dexterous manipulator (SPDM) performing various maintenance tasks on the International Space Station because the real SPDM cannot be physically tested for 3D operations on the ground due to the gravity. The facility uses a high-fidelity SPDM mathematical model, known as the ‘‘truth model’’ of the space robot, to drive a hydraulic robot to mimic the space robot performing contact operations. In this research different techniques were studied for practically verifying that the complex\ud
simulation facility preserves the dynamics of the truth model of the space robot for space-representative contact robotic tasks. Based upon the study and many years of experience in developing and verifying space robotic systems, a practical validation strategy including detailed test cases was developed along with a set of quantitative criteria for judging the validation test results
Reactive Planar Manipulation with Convex Hybrid MPC
This paper presents a reactive controller for planar manipulation tasks that
leverages machine learning to achieve real-time performance. The approach is
based on a Model Predictive Control (MPC) formulation, where the goal is to
find an optimal sequence of robot motions to achieve a desired object motion.
Due to the multiple contact modes associated with frictional interactions, the
resulting optimization program suffers from combinatorial complexity when
tasked with determining the optimal sequence of modes.
To overcome this difficulty, we formulate the search for the optimal mode
sequences offline, separately from the search for optimal control inputs
online. Using tools from machine learning, this leads to a convex hybrid MPC
program that can be solved in real-time. We validate our algorithm on a planar
manipulation experimental setup where results show that the convex hybrid MPC
formulation with learned modes achieves good closed-loop performance on a
trajectory tracking problem
Research and development at ORNL/CESAR towards cooperating robotic systems for hazardous environments
One of the frontiers in intelligent machine research is the understanding of how constructive cooperation among multiple autonomous agents can be effected. The effort at the Center for Engineering Systems Advanced Research (CESAR) at the Oak Ridge National Laboratory (ORNL) focuses on two problem areas: (1) cooperation by multiple mobile robots in dynamic, incompletely known environments; and (2) cooperating robotic manipulators. Particular emphasis is placed on experimental evaluation of research and developments using the CESAR robot system testbeds, including three mobile robots, and a seven-axis, kinematically redundant mobile manipulator. This paper summarizes initial results of research addressing the decoupling of position and force control for two manipulators holding a common object, and the path planning for multiple robots in a common workspace
Neural Network Dynamics for Model-Based Deep Reinforcement Learning with Model-Free Fine-Tuning
Model-free deep reinforcement learning algorithms have been shown to be
capable of learning a wide range of robotic skills, but typically require a
very large number of samples to achieve good performance. Model-based
algorithms, in principle, can provide for much more efficient learning, but
have proven difficult to extend to expressive, high-capacity models such as
deep neural networks. In this work, we demonstrate that medium-sized neural
network models can in fact be combined with model predictive control (MPC) to
achieve excellent sample complexity in a model-based reinforcement learning
algorithm, producing stable and plausible gaits to accomplish various complex
locomotion tasks. We also propose using deep neural network dynamics models to
initialize a model-free learner, in order to combine the sample efficiency of
model-based approaches with the high task-specific performance of model-free
methods. We empirically demonstrate on MuJoCo locomotion tasks that our pure
model-based approach trained on just random action data can follow arbitrary
trajectories with excellent sample efficiency, and that our hybrid algorithm
can accelerate model-free learning on high-speed benchmark tasks, achieving
sample efficiency gains of 3-5x on swimmer, cheetah, hopper, and ant agents.
Videos can be found at https://sites.google.com/view/mbm
Multibody dynamics in robotics with focus on contact events
Multibody dynamics methodologies have been fundamental tools utilized to model and simulate robotic systems that experience contact conditions with the surrounding environment, such as in the case of feet and ground interactions. In addressing such problems, it is of paramount importance to accurately and efficiently handle the large body displacement associated with locomotion of robots, as well as the dynamic response related to contact-impact events. Thus, a generic computational approach, based on the Newton-Euler formulation, to represent the gross motion of robotic systems, is revisited in this work. The main kinematic and dynamic features, necessary to obtain the equations of motion, are discussed. A numerical procedure suitable to solve the equations of motion is also presented. The problem of modeling contacts in dynamical systems involves two main tasks, namely the contact detection and the contact resolution, which take into account for the kinematics and dynamics of the contacting bodies, constituting the general framework for the process of modeling and simulating complex contact scenarios. In order to properly model the contact interactions, the contact kinematic properties are established based on the geometry of contacting bodies, which allow to perform the contact detection task. The contact dynamics is represented by continuous contact force models, both in terms of normal and tangential contact directions. Finally, the presented formulations are demonstrated by the application to several robotics systems that involve contact and impact events with surrounding environment. Special emphasis is put on the systems’ dynamic behavior, in terms of performance and stability
One-Shot Learning of Manipulation Skills with Online Dynamics Adaptation and Neural Network Priors
One of the key challenges in applying reinforcement learning to complex
robotic control tasks is the need to gather large amounts of experience in
order to find an effective policy for the task at hand. Model-based
reinforcement learning can achieve good sample efficiency, but requires the
ability to learn a model of the dynamics that is good enough to learn an
effective policy. In this work, we develop a model-based reinforcement learning
algorithm that combines prior knowledge from previous tasks with online
adaptation of the dynamics model. These two ingredients enable highly
sample-efficient learning even in regimes where estimating the true dynamics is
very difficult, since the online model adaptation allows the method to locally
compensate for unmodeled variation in the dynamics. We encode the prior
experience into a neural network dynamics model, adapt it online by
progressively refitting a local linear model of the dynamics, and use model
predictive control to plan under these dynamics. Our experimental results show
that this approach can be used to solve a variety of complex robotic
manipulation tasks in just a single attempt, using prior data from other
manipulation behaviors
Bipedal Hopping: Reduced-order Model Embedding via Optimization-based Control
This paper presents the design and validation of controlling hopping on the
3D bipedal robot Cassie. A spring-mass model is identified from the kinematics
and compliance of the robot. The spring stiffness and damping are encapsulated
by the leg length, thus actuating the leg length can create and control hopping
behaviors. Trajectory optimization via direct collocation is performed on the
spring-mass model to plan jumping and landing motions. The leg length
trajectories are utilized as desired outputs to synthesize a control Lyapunov
function based quadratic program (CLF-QP). Centroidal angular momentum, taking
as an addition output in the CLF-QP, is also stabilized in the jumping phase to
prevent whole body rotation in the underactuated flight phase. The solution to
the CLF-QP is a nonlinear feedback control law that achieves dynamic jumping
behaviors on bipedal robots with compliance. The framework presented in this
paper is verified experimentally on the bipedal robot Cassie.Comment: 8 pages, 7 figures, accepted by IROS 201
- …