143 research outputs found
Fast Damage Recovery in Robotics with the T-Resilience Algorithm
Damage recovery is critical for autonomous robots that need to operate for a
long time without assistance. Most current methods are complex and costly
because they require anticipating each potential damage in order to have a
contingency plan ready. As an alternative, we introduce the T-resilience
algorithm, a new algorithm that allows robots to quickly and autonomously
discover compensatory behaviors in unanticipated situations. This algorithm
equips the robot with a self-model and discovers new behaviors by learning to
avoid those that perform differently in the self-model and in reality. Our
algorithm thus does not identify the damaged parts but it implicitly searches
for efficient behaviors that do not use them. We evaluate the T-Resilience
algorithm on a hexapod robot that needs to adapt to leg removal, broken legs
and motor failures; we compare it to stochastic local search, policy gradient
and the self-modeling algorithm proposed by Bongard et al. The behavior of the
robot is assessed on-board thanks to a RGB-D sensor and a SLAM algorithm. Using
only 25 tests on the robot and an overall running time of 20 minutes,
T-Resilience consistently leads to substantially better results than the other
approaches
Robust Quadrupedal Locomotion via Risk-Averse Policy Learning
The robustness of legged locomotion is crucial for quadrupedal robots in
challenging terrains. Recently, Reinforcement Learning (RL) has shown promising
results in legged locomotion and various methods try to integrate privileged
distillation, scene modeling, and external sensors to improve the
generalization and robustness of locomotion policies. However, these methods
are hard to handle uncertain scenarios such as abrupt terrain changes or
unexpected external forces. In this paper, we consider a novel risk-sensitive
perspective to enhance the robustness of legged locomotion. Specifically, we
employ a distributional value function learned by quantile regression to model
the aleatoric uncertainty of environments, and perform risk-averse policy
learning by optimizing the worst-case scenarios via a risk distortion measure.
Extensive experiments in both simulation environments and a real Aliengo robot
demonstrate that our method is efficient in handling various external
disturbances, and the resulting policy exhibits improved robustness in harsh
and uncertain situations in legged locomotion. Videos are available at
https://risk-averse-locomotion.github.io/.Comment: 8 pages, 5 figure
Locomotion gait optimization for a quadruped robot
This article describes the development of a gait optimization
system that allows a fast but stable robot quadruped crawl gait.
We focus in the development of a quadruped robot walking gait
locomotion that combine bio-inspired Central Patterns Generators
(CPGs) and Genetic Algorithms (GA). The CPGs are modelled as
autonomous differential equations, that generate the necessary limb
movement to perform the walking gait, and the Genetic Algorithm
perform the search of the CPGs parameters.
This approach allows to explicitly specify parameters such as amplitude,
offset and frequency of movement and to smoothly modulate
the generated trajectories according to changes in these parameters.
It is therefore easy to combine the CPG with an optimization method.
A genetic algorithm determines the best set of parameters that generates
the limbs movements. We intend to obtain a walking gait locomotion
that minimizes the vibration and maximizes the wide stability
margin and the forward velocity.
The experimental results, performed on a simulated Aibo robot,
demonstrated that our approach allows low vibration with a high velocity
and wide stability margin for a quadruped walking gait locomotion
Legged Robots for Object Manipulation: A Review
Legged robots can have a unique role in manipulating objects in dynamic,
human-centric, or otherwise inaccessible environments. Although most legged
robotics research to date typically focuses on traversing these challenging
environments, many legged platform demonstrations have also included "moving an
object" as a way of doing tangible work. Legged robots can be designed to
manipulate a particular type of object (e.g., a cardboard box, a soccer ball,
or a larger piece of furniture), by themselves or collaboratively. The
objective of this review is to collect and learn from these examples, to both
organize the work done so far in the community and highlight interesting open
avenues for future work. This review categorizes existing works into four main
manipulation methods: object interactions without grasping, manipulation with
walking legs, dedicated non-locomotive arms, and legged teams. Each method has
different design and autonomy features, which are illustrated by available
examples in the literature. Based on a few simplifying assumptions, we further
provide quantitative comparisons for the range of possible relative sizes of
the manipulated object with respect to the robot. Taken together, these
examples suggest new directions for research in legged robot manipulation, such
as multifunctional limbs, terrain modeling, or learning-based control, to
support a number of new deployments in challenging indoor/outdoor scenarios in
warehouses/construction sites, preserved natural areas, and especially for home
robotics.Comment: Preprint of the paper submitted to Frontiers in Mechanical
Engineerin
Learning Complex Motor Skills for Legged Robot Fall Recovery
Falling is inevitable for legged robots in challenging real-world scenarios, where environments are unstructured and situations are unpredictable, such as uneven terrain in the wild. Hence, to recover from falls and achieve all-terrain traversability, it is essential for intelligent robots to possess the complex motor skills required to resume operation. To go beyond the limitation of handcrafted control, we investigated a deep reinforcement learning approach to learn generalized feedback-control policies for fall recovery that are robust to external disturbances. We proposed a design guideline for selecting key states for initialization, including a comparison to the random state initialization. The proposed learning-based pipeline is applicable to different robot models and their corner cases, including both small-/large-size bipeds and quadrupeds. Further, we show that the learned fall recovery policies are hardware-feasible and can be implemented on real robots
Multi-Objective Optimization for Speed and Stability of a Sony Aibo Gait
Locomotion is a fundamental facet of mobile robotics that many higher level aspects rely on. However, this is not a simple problem for legged robots with many degrees of freedom. For this reason, machine learning techniques have been applied to the domain. Although impressive results have been achieved, there remains a fundamental problem with using most machine learning methods. The learning algorithms usually require a large dataset which is prohibitively hard to collect on an actual robot. Further, learning in simulation has had limited success transitioning to the real world. Also, many learning algorithms optimize for a single fitness function, neglecting many of the effects on other parts of the system. As part of the RoboCup 4-legged league, many researchers have worked on increasing the walking/gait speed of Sony AIBO robots. Recently, the effort shifted from developing a quick gait, to developing a gait that also provides a stable sensing platform. However, to date, optimization of both velocity and camera stability has only occurred using a single fitness function that incorporates the two objectives with a weighting that defines the desired tradeoff between them. However, the true nature of this tradeoff is not understood because the pareto front has never been charted, so this a priori decision is uninformed. This project applies the Nondominated Sorting Genetic Algorithm-II (NSGA-II) to find a pareto set of fast, stable gait parameters. This allows a user to select the best tradeoff between balance and speed for a given application. Three fitness functions are defined: one speed measure and two stability measures. A plot of evolved gaits shows a pareto front that indicates speed and stability are indeed conflicting goals. Interestingly, the results also show that tradeoffs also exist between different measures of stability
Creating a Dynamic Quadrupedal Robotic Goalkeeper with Reinforcement Learning
We present a reinforcement learning (RL) framework that enables quadrupedal
robots to perform soccer goalkeeping tasks in the real world. Soccer
goalkeeping using quadrupeds is a challenging problem, that combines highly
dynamic locomotion with precise and fast non-prehensile object (ball)
manipulation. The robot needs to react to and intercept a potentially flying
ball using dynamic locomotion maneuvers in a very short amount of time, usually
less than one second. In this paper, we propose to address this problem using a
hierarchical model-free RL framework. The first component of the framework
contains multiple control policies for distinct locomotion skills, which can be
used to cover different regions of the goal. Each control policy enables the
robot to track random parametric end-effector trajectories while performing one
specific locomotion skill, such as jump, dive, and sidestep. These skills are
then utilized by the second part of the framework which is a high-level planner
to determine a desired skill and end-effector trajectory in order to intercept
a ball flying to different regions of the goal. We deploy the proposed
framework on a Mini Cheetah quadrupedal robot and demonstrate the effectiveness
of our framework for various agile interceptions of a fast-moving ball in the
real world.Comment: First two authors contributed equally. Accompanying video is at
https://youtu.be/iX6OgG67-Z
- …