1,827 research outputs found
Robust Quadrupedal Locomotion via Risk-Averse Policy Learning
The robustness of legged locomotion is crucial for quadrupedal robots in
challenging terrains. Recently, Reinforcement Learning (RL) has shown promising
results in legged locomotion and various methods try to integrate privileged
distillation, scene modeling, and external sensors to improve the
generalization and robustness of locomotion policies. However, these methods
are hard to handle uncertain scenarios such as abrupt terrain changes or
unexpected external forces. In this paper, we consider a novel risk-sensitive
perspective to enhance the robustness of legged locomotion. Specifically, we
employ a distributional value function learned by quantile regression to model
the aleatoric uncertainty of environments, and perform risk-averse policy
learning by optimizing the worst-case scenarios via a risk distortion measure.
Extensive experiments in both simulation environments and a real Aliengo robot
demonstrate that our method is efficient in handling various external
disturbances, and the resulting policy exhibits improved robustness in harsh
and uncertain situations in legged locomotion. Videos are available at
https://risk-averse-locomotion.github.io/.Comment: 8 pages, 5 figure
HumanMimic: Learning Natural Locomotion and Transitions for Humanoid Robot via Wasserstein Adversarial Imitation
Transferring human motion skills to humanoid robots remains a significant
challenge. In this study, we introduce a Wasserstein adversarial imitation
learning system, allowing humanoid robots to replicate natural whole-body
locomotion patterns and execute seamless transitions by mimicking human
motions. First, we present a unified primitive-skeleton motion retargeting to
mitigate morphological differences between arbitrary human demonstrators and
humanoid robots. An adversarial critic component is integrated with
Reinforcement Learning (RL) to guide the control policy to produce behaviors
aligned with the data distribution of mixed reference motions. Additionally, we
employ a specific Integral Probabilistic Metric (IPM), namely the Wasserstein-1
distance with a novel soft boundary constraint to stabilize the training
process and prevent model collapse. Our system is evaluated on a full-sized
humanoid JAXON in the simulator. The resulting control policy demonstrates a
wide range of locomotion patterns, including standing, push-recovery, squat
walking, human-like straight-leg walking, and dynamic running. Notably, even in
the absence of transition motions in the demonstration dataset, robots showcase
an emerging ability to transit naturally between distinct locomotion patterns
as desired speed changes
Learning Agility and Adaptive Legged Locomotion via Curricular Hindsight Reinforcement Learning
Agile and adaptive maneuvers such as fall recovery, high-speed turning, and
sprinting in the wild are challenging for legged systems. We propose a
Curricular Hindsight Reinforcement Learning (CHRL) that learns an end-to-end
tracking controller that achieves powerful agility and adaptation for the
legged robot. The two key components are (I) a novel automatic curriculum
strategy on task difficulty and (ii) a Hindsight Experience Replay strategy
adapted to legged locomotion tasks. We demonstrated successful agile and
adaptive locomotion on a real quadruped robot that performed fall recovery
autonomously, coherent trotting, sustained outdoor speeds up to 3.45 m/s, and
tuning speeds up to 3.2 rad/s. This system produces adaptive behaviours
responding to changing situations and unexpected disturbances on natural
terrains like grass and dirt
- …