784 research outputs found
3LP: a linear 3D-walking model including torso and swing dynamics
In this paper, we present a new model of biped locomotion which is composed
of three linear pendulums (one per leg and one for the whole upper body) to
describe stance, swing and torso dynamics. In addition to double support, this
model has different actuation possibilities in the swing hip and stance ankle
which could be widely used to produce different walking gaits. Without the need
for numerical time-integration, closed-form solutions help finding periodic
gaits which could be simply scaled in certain dimensions to modulate the motion
online. Thanks to linearity properties, the proposed model can provide a
computationally fast platform for model predictive controllers to predict the
future and consider meaningful inequality constraints to ensure feasibility of
the motion. Such property is coming from describing dynamics with joint torques
directly and therefore, reflecting hardware limitations more precisely, even in
the very abstract high level template space. The proposed model produces
human-like torque and ground reaction force profiles and thus, compared to
point-mass models, it is more promising for precise control of humanoid robots.
Despite being linear and lacking many other features of human walking like CoM
excursion, knee flexion and ground clearance, we show that the proposed model
can predict one of the main optimality trends in human walking, i.e. nonlinear
speed-frequency relationship. In this paper, we mainly focus on describing the
model and its capabilities, comparing it with human data and calculating
optimal human gait variables. Setting up control problems and advanced
biomechanical analysis still remain for future works.Comment: Journal paper under revie
Push recovery with stepping strategy based on time-projection control
In this paper, we present a simple control framework for on-line push
recovery with dynamic stepping properties. Due to relatively heavy legs in our
robot, we need to take swing dynamics into account and thus use a linear model
called 3LP which is composed of three pendulums to simulate swing and torso
dynamics. Based on 3LP equations, we formulate discrete LQR controllers and use
a particular time-projection method to adjust the next footstep location
on-line during the motion continuously. This adjustment, which is found based
on both pelvis and swing foot tracking errors, naturally takes the swing
dynamics into account. Suggested adjustments are added to the Cartesian 3LP
gaits and converted to joint-space trajectories through inverse kinematics.
Fixed and adaptive foot lift strategies also ensure enough ground clearance in
perturbed walking conditions. The proposed structure is robust, yet uses very
simple state estimation and basic position tracking. We rely on the physical
series elastic actuators to absorb impacts while introducing simple laws to
compensate their tracking bias. Extensive experiments demonstrate the
functionality of different control blocks and prove the effectiveness of
time-projection in extreme push recovery scenarios. We also show self-produced
and emergent walking gaits when the robot is subject to continuous dragging
forces. These gaits feature dynamic walking robustness due to relatively soft
springs in the ankles and avoiding any Zero Moment Point (ZMP) control in our
proposed architecture.Comment: 20 pages journal pape
Beyond Basins of Attraction: Quantifying Robustness of Natural Dynamics
Properly designing a system to exhibit favorable natural dynamics can greatly
simplify designing or learning the control policy. However, it is still unclear
what constitutes favorable natural dynamics and how to quantify its effect.
Most studies of simple walking and running models have focused on the basins of
attraction of passive limit-cycles and the notion of self-stability. We instead
emphasize the importance of stepping beyond basins of attraction. We show an
approach based on viability theory to quantify robust sets in state-action
space. These sets are valid for the family of all robust control policies,
which allows us to quantify the robustness inherent to the natural dynamics
before designing the control policy or specifying a control objective. We
illustrate our formulation using spring-mass models, simple low dimensional
models of running systems. We then show an example application by optimizing
robustness of a simulated planar monoped, using a gradient-free optimization
scheme. Both case studies result in a nonlinear effective stiffness providing
more robustness.Comment: 15 pages. This work has been accepted to IEEE Transactions on
Robotics (2019
An Efficiently Solvable Quadratic Program for Stabilizing Dynamic Locomotion
We describe a whole-body dynamic walking controller implemented as a convex
quadratic program. The controller solves an optimal control problem using an
approximate value function derived from a simple walking model while respecting
the dynamic, input, and contact constraints of the full robot dynamics. By
exploiting sparsity and temporal structure in the optimization with a custom
active-set algorithm, we surpass the performance of the best available
off-the-shelf solvers and achieve 1kHz control rates for a 34-DOF humanoid. We
describe applications to balancing and walking tasks using the simulated Atlas
robot in the DARPA Virtual Robotics Challenge.Comment: 6 pages, published at ICRA 201
From Knowing to Doing: Learning Diverse Motor Skills through Instruction Learning
Recent years have witnessed many successful trials in the robot learning
field. For contact-rich robotic tasks, it is challenging to learn coordinated
motor skills by reinforcement learning. Imitation learning solves this problem
by using a mimic reward to encourage the robot to track a given reference
trajectory. However, imitation learning is not so efficient and may constrain
the learned motion. In this paper, we propose instruction learning, which is
inspired by the human learning process and is highly efficient, flexible, and
versatile for robot motion learning. Instead of using a reference signal in the
reward, instruction learning applies a reference signal directly as a
feedforward action, and it is combined with a feedback action learned by
reinforcement learning to control the robot. Besides, we propose the action
bounding technique and remove the mimic reward, which is shown to be crucial
for efficient and flexible learning. We compare the performance of instruction
learning with imitation learning, indicating that instruction learning can
greatly speed up the training process and guarantee learning the desired motion
correctly. The effectiveness of instruction learning is validated through a
bunch of motion learning examples for a biped robot and a quadruped robot,
where skills can be learned typically within several million steps. Besides, we
also conduct sim-to-real transfer and online learning experiments on a real
quadruped robot. Instruction learning has shown great merits and potential,
making it a promising alternative for imitation learning
- …