1,003 research outputs found
Safe trajectory optimization for whole-body motion of humanoids
International audienceMulti-task prioritized controllers generate complex behaviors for humanoids that concurrently satisfy several tasks and constraints. In our previous work we automatically learned the task priorities that maximized the robot performance in whole-body reaching tasks, ensuring that the optimized priorities were leading to safe behaviors. Here, we take the opposite approach: we optimize the task trajectories for whole-body balancing tasks with switching contacts, ensuring that the optimized movements are safe and never violate any of the robot and problem constraints. We use (1+1)-CMA-ES with Constrained Covariance Adaptation as a constrained black box stochastic optimization algorithm, with an instance of (1+1)-CMA-ES for bootstrapping the search. We apply our learning framework to the prioritized whole-body torque controller of iCub, to optimize the robot's movement for standing up from a chair
Whole-Body MPC for a Dynamically Stable Mobile Manipulator
Autonomous mobile manipulation offers a dual advantage of mobility provided
by a mobile platform and dexterity afforded by the manipulator. In this paper,
we present a whole-body optimal control framework to jointly solve the problems
of manipulation, balancing and interaction as one optimization problem for an
inherently unstable robot. The optimization is performed using a Model
Predictive Control (MPC) approach; the optimal control problem is transcribed
at the end-effector space, treating the position and orientation tasks in the
MPC planner, and skillfully planning for end-effector contact forces. The
proposed formulation evaluates how the control decisions aimed at end-effector
tracking and environment interaction will affect the balance of the system in
the future. We showcase the advantages of the proposed MPC approach on the
example of a ball-balancing robot with a robotic manipulator and validate our
controller in hardware experiments for tasks such as end-effector pose tracking
and door opening
A framework for safe human-humanoid coexistence
This work is focused on the development of a safety framework for Human-Humanoid coexistence, with emphasis on humanoid locomotion. After a brief introduction to the fundamental concepts of humanoid locomotion, the two most common approaches for gait generation are presented, and are extended with the inclusion of a stability condition to guarantee the boundedness of the generated trajectories. Then the safety framework is presented, with the introduction of different safety behaviors. These behaviors are meant to enhance the overall level of safety during any robot operation. Proactive behaviors will enhance or adapt the current robot operations to reduce the risk of danger, while override behaviors will stop the current robot activity in order to take action against a particularly dangerous situation. A state
machine is defined to control the transitions between the behaviors. The behaviors that are strictly related to locomotion are subsequently detailed, and an implementation is proposed and validated. A possible implementation of the remaining behaviors is proposed through the review of related works that can be found in literature
Learning a Unified Control Policy for Safe Falling
Being able to fall safely is a necessary motor skill for humanoids performing
highly dynamic tasks, such as running and jumping. We propose a new method to
learn a policy that minimizes the maximal impulse during the fall. The
optimization solves for both a discrete contact planning problem and a
continuous optimal control problem. Once trained, the policy can compute the
optimal next contacting body part (e.g. left foot, right foot, or hands),
contact location and timing, and the required joint actuation. We represent the
policy as a mixture of actor-critic neural network, which consists of n control
policies and the corresponding value functions. Each pair of actor-critic is
associated with one of the n possible contacting body parts. During execution,
the policy corresponding to the highest value function will be executed while
the associated body part will be the next contact with the ground. With this
mixture of actor-critic architecture, the discrete contact sequence planning is
solved through the selection of the best critics while the continuous control
problem is solved by the optimization of actors. We show that our policy can
achieve comparable, sometimes even higher, rewards than a recursive search of
the action space using dynamic programming, while enjoying 50 to 400 times of
speed gain during online execution
- …