48 research outputs found
Deep Reinforcement Learning for Tensegrity Robot Locomotion
Tensegrity robots, composed of rigid rods connected by elastic cables, have a
number of unique properties that make them appealing for use as planetary
exploration rovers. However, control of tensegrity robots remains a difficult
problem due to their unusual structures and complex dynamics. In this work, we
show how locomotion gaits can be learned automatically using a novel extension
of mirror descent guided policy search (MDGPS) applied to periodic locomotion
movements, and we demonstrate the effectiveness of our approach on tensegrity
robot locomotion. We evaluate our method with real-world and simulated
experiments on the SUPERball tensegrity robot, showing that the learned
policies generalize to changes in system parameters, unreliable sensor
measurements, and variation in environmental conditions, including varied
terrains and a range of different gravities. Our experiments demonstrate that
our method not only learns fast, power-efficient feedback policies for rolling
gaits, but that these policies can succeed with only the limited onboard
sensing provided by SUPERball's accelerometers. We compare the learned feedback
policies to learned open-loop policies and hand-engineered controllers, and
demonstrate that the learned policy enables the first continuous, reliable
locomotion gait for the real SUPERball robot. Our code and other supplementary
materials are available from http://rll.berkeley.edu/drl_tensegrityComment: International Conference on Robotics and Automation (ICRA), 2017.
Project website link is http://rll.berkeley.edu/drl_tensegrit
Exploring the Behavior Repertoire of a Wireless Vibrationally Actuated Tensegrity Robot
Soft robotics is an emerging field of research due to its potential to explore and operate in unstructured, rugged, and dynamic environments. However, the properties that make soft robots compelling also make them difficult to robustly control. Here at Union, we developed the world’s first wireless soft tensegrity robot. The goal of my thesis is to explore effective and efficient methods to explore the diverse behavior our tensegrity robot. We will achieve that by applying state-of-art machine learning technique and a novelty search algorithm
Real2Sim2Real Transfer for Control of Cable-driven Robots via a Differentiable Physics Engine
Tensegrity robots, composed of rigid rods and flexible cables, exhibit high
strength-to-weight ratios and extreme deformations, enabling them to navigate
unstructured terrain and even survive harsh impacts. However, they are hard to
control due to their high dimensionality, complex dynamics, and coupled
architecture. Physics-based simulation is one avenue for developing locomotion
policies that can then be transferred to real robots, but modeling tensegrity
robots is a complex task, so simulations experience a substantial sim2real gap.
To address this issue, this paper describes a Real2Sim2Real strategy for
tensegrity robots. This strategy is based on a differential physics engine that
can be trained given limited data from a real robot (i.e. offline measurements
and one random trajectory) and achieve a high enough accuracy to discover
transferable locomotion policies. Beyond the overall pipeline, key
contributions of this work include computing non-zero gradients at contact
points, a loss function, and a trajectory segmentation technique that avoid
conflicts in gradient evaluation during training. The proposed pipeline is
demonstrated and evaluated on a real 3-bar tensegrity robot.Comment: Submitted to ICRA202
Optimizing Tensegrity Gaits Using Bayesian Optimization
We design and implement a new, modular, more complex tensegrity robot featuring data collection and wireless communication and operation as well as necessary accompanying research infrastructure. We then utilize this new tensegrity to assess previous research on using Bayesian optimization to generate effective forward gaits for tensegrity robots. Ultimately, we affirm the conclusions of previous researchers, demonstrating that Bayesian optimization is statistically significantly (p \u3c 0:05) more effective at discovering useful gaits than random search. We also identify several flaws in our new system and identify means of addressing them, paving the way for more effective future research
Opinions and Outlooks on Morphological Computation
Morphological Computation is based on the observation that biological systems seem to carry out relevant computations with their morphology (physical body) in order to successfully interact with their environments. This can be observed in a whole range of systems and at many different scales. It has been studied in animals – e.g., while running, the functionality of coping with impact and slight unevenness in the ground is "delivered" by the shape of the legs and the damped elasticity of the muscle-tendon system – and plants, but it has also been observed at the cellular and even at the molecular level – as seen, for example, in spontaneous self-assembly. The concept of morphological computation has served as an inspirational resource to build bio-inspired robots, design novel approaches for support systems in health care, implement computation with natural systems, but also in art and architecture. As a consequence, the field is highly interdisciplinary, which is also nicely reflected in the wide range of authors that are featured in this e-book. We have contributions from robotics, mechanical engineering, health, architecture, biology, philosophy, and others
A survey on policy search algorithms for learning robot controllers in a handful of trials
Most policy search algorithms require thousands of training episodes to find
an effective policy, which is often infeasible with a physical robot. This
survey article focuses on the extreme other end of the spectrum: how can a
robot adapt with only a handful of trials (a dozen) and a few minutes? By
analogy with the word "big-data", we refer to this challenge as "micro-data
reinforcement learning". We show that a first strategy is to leverage prior
knowledge on the policy structure (e.g., dynamic movement primitives), on the
policy parameters (e.g., demonstrations), or on the dynamics (e.g.,
simulators). A second strategy is to create data-driven surrogate models of the
expected reward (e.g., Bayesian optimization) or the dynamical model (e.g.,
model-based policy search), so that the policy optimizer queries the model
instead of the real system. Overall, all successful micro-data algorithms
combine these two strategies by varying the kind of model and prior knowledge.
The current scientific challenges essentially revolve around scaling up to
complex robots (e.g., humanoids), designing generic priors, and optimizing the
computing time.Comment: 21 pages, 3 figures, 4 algorithms, accepted at IEEE Transactions on
Robotic