182 research outputs found
The Director: A Composable Behaviour System with Soft Transitions
Software frameworks for behaviour are critical in robotics as they enable the
correct and efficient execution of functions. While modern behaviour systems
have improved their composability, they do not focus on smooth transitions and
often lack functionality. In this work, we present the Director, a novel
behaviour framework and algorithm that addresses these problems. It has
functionality for soft transitions, multiple implementations of the same action
chosen based on conditionals, and strict resource control. This system has
shown success in the Humanoid Kid Size 2022/2023 Virtual Season and the
Humanoid Kid Size RoboCup 2023 Bordeaux competition
Legged Robots for Object Manipulation: A Review
Legged robots can have a unique role in manipulating objects in dynamic,
human-centric, or otherwise inaccessible environments. Although most legged
robotics research to date typically focuses on traversing these challenging
environments, many legged platform demonstrations have also included "moving an
object" as a way of doing tangible work. Legged robots can be designed to
manipulate a particular type of object (e.g., a cardboard box, a soccer ball,
or a larger piece of furniture), by themselves or collaboratively. The
objective of this review is to collect and learn from these examples, to both
organize the work done so far in the community and highlight interesting open
avenues for future work. This review categorizes existing works into four main
manipulation methods: object interactions without grasping, manipulation with
walking legs, dedicated non-locomotive arms, and legged teams. Each method has
different design and autonomy features, which are illustrated by available
examples in the literature. Based on a few simplifying assumptions, we further
provide quantitative comparisons for the range of possible relative sizes of
the manipulated object with respect to the robot. Taken together, these
examples suggest new directions for research in legged robot manipulation, such
as multifunctional limbs, terrain modeling, or learning-based control, to
support a number of new deployments in challenging indoor/outdoor scenarios in
warehouses/construction sites, preserved natural areas, and especially for home
robotics.Comment: Preprint of the paper submitted to Frontiers in Mechanical
Engineerin
Hierarchical Reinforcement Learning for Precise Soccer Shooting Skills using a Quadrupedal Robot
We address the problem of enabling quadrupedal robots to perform precise
shooting skills in the real world using reinforcement learning. Developing
algorithms to enable a legged robot to shoot a soccer ball to a given target is
a challenging problem that combines robot motion control and planning into one
task. To solve this problem, we need to consider the dynamics limitation and
motion stability during the control of a dynamic legged robot. Moreover, we
need to consider motion planning to shoot the hard-to-model deformable ball
rolling on the ground with uncertain friction to a desired location. In this
paper, we propose a hierarchical framework that leverages deep reinforcement
learning to train (a) a robust motion control policy that can track arbitrary
motions and (b) a planning policy to decide the desired kicking motion to shoot
a soccer ball to a target. We deploy the proposed framework on an A1
quadrupedal robot and enable it to accurately shoot the ball to random targets
in the real world.Comment: Accepted to 2022 IEEE/RSJ International Conference on Intelligent
Robots and Systems (IROS 2022
Scaled Autonomy for Networked Humanoids
Humanoid robots have been developed with the intention of aiding in environments designed for humans. As such, the control of humanoid morphology and effectiveness of human robot interaction form the two principal research issues for deploying these robots in the real world. In this thesis work, the issue of humanoid control is coupled with human robot interaction under the framework of scaled autonomy, where the human and robot exchange levels of control depending on the environment and task at hand. This scaled autonomy is approached with control algorithms for reactive stabilization of human commands and planned trajectories that encode semantically meaningful motion preferences in a sequential convex optimization framework.
The control and planning algorithms have been extensively tested in the field for robustness and system verification. The RoboCup competition provides a benchmark competition for autonomous agents that are trained with a human supervisor. The kid-sized and adult-sized humanoid robots coordinate over a noisy network in a known environment with adversarial opponents, and the software and routines in this work allowed for five consecutive championships. Furthermore, the motion planning and user interfaces developed in the work have been tested in the noisy network of the DARPA Robotics Challenge (DRC) Trials and Finals in an unknown environment.
Overall, the ability to extend simplified locomotion models to aid in semi-autonomous manipulation allows untrained humans to operate complex, high dimensional robots. This represents another step in the path to deploying humanoids in the real world, based on the low dimensional motion abstractions and proven performance in real world tasks like RoboCup and the DRC
Learning Agile Soccer Skills for a Bipedal Robot with Deep Reinforcement Learning
We investigate whether Deep Reinforcement Learning (Deep RL) is able to
synthesize sophisticated and safe movement skills for a low-cost, miniature
humanoid robot that can be composed into complex behavioral strategies in
dynamic environments. We used Deep RL to train a humanoid robot with 20
actuated joints to play a simplified one-versus-one (1v1) soccer game. We first
trained individual skills in isolation and then composed those skills
end-to-end in a self-play setting. The resulting policy exhibits robust and
dynamic movement skills such as rapid fall recovery, walking, turning, kicking
and more; and transitions between them in a smooth, stable, and efficient
manner - well beyond what is intuitively expected from the robot. The agents
also developed a basic strategic understanding of the game, and learned, for
instance, to anticipate ball movements and to block opponent shots. The full
range of behaviors emerged from a small set of simple rewards. Our agents were
trained in simulation and transferred to real robots zero-shot. We found that a
combination of sufficiently high-frequency control, targeted dynamics
randomization, and perturbations during training in simulation enabled
good-quality transfer, despite significant unmodeled effects and variations
across robot instances. Although the robots are inherently fragile, minor
hardware modifications together with basic regularization of the behavior
during training led the robots to learn safe and effective movements while
still performing in a dynamic and agile way. Indeed, even though the agents
were optimized for scoring, in experiments they walked 156% faster, took 63%
less time to get up, and kicked 24% faster than a scripted baseline, while
efficiently combining the skills to achieve the longer term objectives.
Examples of the emergent behaviors and full 1v1 matches are available on the
supplementary website.Comment: Project website: https://sites.google.com/view/op3-socce
Locomoção de humanoides robusta e versátil baseada em controlo analĂtico e fĂsica residual
Humanoid robots are made to resemble humans but their locomotion
abilities are far from ours in terms of agility and versatility. When humans
walk on complex terrains or face external disturbances, they
combine a set of strategies, unconsciously and efficiently, to regain
stability. This thesis tackles the problem of developing a robust omnidirectional
walking framework, which is able to generate versatile
and agile locomotion on complex terrains. We designed and developed
model-based and model-free walk engines and formulated the
controllers using different approaches including classical and optimal
control schemes and validated their performance through simulations
and experiments. These frameworks have hierarchical structures that
are composed of several layers. These layers are composed of several
modules that are connected together to fade the complexity and
increase the flexibility of the proposed frameworks. Additionally, they
can be easily and quickly deployed on different platforms.
Besides, we believe that using machine learning on top of analytical approaches
is a key to open doors for humanoid robots to step out of laboratories.
We proposed a tight coupling between analytical control and
deep reinforcement learning. We augmented our analytical controller
with reinforcement learning modules to learn how to regulate the walk
engine parameters (planners and controllers) adaptively and generate
residuals to adjust the robot’s target joint positions (residual physics).
The effectiveness of the proposed frameworks was demonstrated and
evaluated across a set of challenging simulation scenarios. The robot
was able to generalize what it learned in one scenario, by displaying
human-like locomotion skills in unforeseen circumstances, even in the
presence of noise and external pushes.Os robĂ´s humanoides sĂŁo feitos para se parecerem com humanos,
mas suas habilidades de locomoção estão longe das nossas em termos
de agilidade e versatilidade. Quando os humanos caminham em
terrenos complexos ou enfrentam distĂşrbios externos combinam diferentes
estratégias, de forma inconsciente e eficiente, para recuperar a
estabilidade. Esta tese aborda o problema de desenvolver um sistema
robusto para andar de forma omnidirecional, capaz de gerar uma locomoção
para robôs humanoides versátil e ágil em terrenos complexos.
Projetámos e desenvolvemos motores de locomoção sem modelos e
baseados em modelos. Formulámos os controladores usando diferentes
abordagens, incluindo esquemas de controlo clássicos e ideais,
e validámos o seu desempenho por meio de simulações e experiências
reais. Estes frameworks têm estruturas hierárquicas compostas por
várias camadas. Essas camadas são compostas por vários módulos
que sĂŁo conectados entre si para diminuir a complexidade e aumentar
a flexibilidade dos frameworks propostos. Adicionalmente, o sistema
pode ser implementado em diferentes plataformas de forma fácil.
Acreditamos que o uso de aprendizagem automática sobre abordagens
analĂticas Ă© a chave para abrir as portas para robĂ´s humanoides
saĂrem dos laboratĂłrios. Propusemos um forte acoplamento entre controlo
analĂtico e aprendizagem profunda por reforço. Expandimos o
nosso controlador analĂtico com mĂłdulos de aprendizagem por reforço
para aprender como regular os parâmetros do motor de caminhada
(planeadores e controladores) de forma adaptativa e gerar resĂduos
para ajustar as posições das juntas alvo do robĂ´ (fĂsica residual). A
eficácia das estruturas propostas foi demonstrada e avaliada em um
conjunto de cenários de simulação desafiadores. O robô foi capaz de
generalizar o que aprendeu em um cenário, exibindo habilidades de
locomoção humanas em circunstâncias imprevistas, mesmo na presença
de ruĂdo e impulsos externos.Programa Doutoral em Informátic
- …