2,727 research outputs found

    Comparing trotting and turning strategies on the quadrupedal Oncilla Robot

    Get PDF
    In this paper, we compare three different trotting techniques and five different turning strategies on a small, compliant, biologically inspired quadrupedal robot, the Oncilla. The locomotion techniques were optimized on the actual hardware using a treadmill setup, without relying on models. We found that using half ellipses as foot trajectories resulted in the fastest gaits, as well as the highest robustness against parameter changes. Furthermore, we analyzed the importance of using the scapulae for turning, from which we observed that although not necessary, they are needed for turning with a higher speed

    Fingerprint Policy Optimisation for Robust Reinforcement Learning

    Full text link
    Policy gradient methods ignore the potential value of adjusting environment variables: unobservable state features that are randomly determined by the environment in a physical setting, but are controllable in a simulator. This can lead to slow learning, or convergence to suboptimal policies, if the environment variable has a large impact on the transition dynamics. In this paper, we present fingerprint policy optimisation (FPO), which finds a policy that is optimal in expectation across the distribution of environment variables. The central idea is to use Bayesian optimisation (BO) to actively select the distribution of the environment variable that maximises the improvement generated by each iteration of the policy gradient method. To make this BO practical, we contribute two easy-to-compute low-dimensional fingerprints of the current policy. Our experiments show that FPO can efficiently learn policies that are robust to significant rare events, which are unlikely to be observable under random sampling, but are key to learning good policies.Comment: ICML 201

    Intelligent approaches in locomotion - a review

    Get PDF

    A Benchmarking of DCM Based Architectures for Position and Velocity Controlled Walking of Humanoid Robots

    Full text link
    This paper contributes towards the development and comparison of Divergent-Component-of-Motion (DCM) based control architectures for humanoid robot locomotion. More precisely, we present and compare several DCM based implementations of a three layer control architecture. From top to bottom, these three layers are here called: trajectory optimization, simplified model control, and whole-body QP control. All layers use the DCM concept to generate references for the layer below. For the simplified model control layer, we present and compare both instantaneous and Receding Horizon Control controllers. For the whole-body QP control layer, we present and compare controllers for position and velocity control robots. Experimental results are carried out on the one-meter tall iCub humanoid robot. We show which implementation of the above control architecture allows the robot to achieve a walking velocity of 0.41 meters per second.Comment: Submitted to Humanoids201
    corecore