8,923 research outputs found

    Bearing-only formation control with auxiliary distance measurements, leaders, and collision avoidance

    Full text link
    We address the controller synthesis problem for distributed formation control. Our solution requires only relative bearing measurements (as opposed to full translations), and is based on the exact gradient of a Lyapunov function with only global minimizers (independently from the formation topology). These properties allow a simple proof of global asymptotic convergence, and extensions for including distance measurements, leaders and collision avoidance. We validate our approach through simulations and comparison with other stateof-the-art algorithms.ARL grant W911NF-08-2-0004, ARO grant W911NF-13-1-0350, ONR grants N00014-07-1-0829, N00014-14-1-0510, N00014-15-1-2115, NSF grant IIS-1426840, CNS-1521617 and United Technologies

    Whole-Body MPC for a Dynamically Stable Mobile Manipulator

    Full text link
    Autonomous mobile manipulation offers a dual advantage of mobility provided by a mobile platform and dexterity afforded by the manipulator. In this paper, we present a whole-body optimal control framework to jointly solve the problems of manipulation, balancing and interaction as one optimization problem for an inherently unstable robot. The optimization is performed using a Model Predictive Control (MPC) approach; the optimal control problem is transcribed at the end-effector space, treating the position and orientation tasks in the MPC planner, and skillfully planning for end-effector contact forces. The proposed formulation evaluates how the control decisions aimed at end-effector tracking and environment interaction will affect the balance of the system in the future. We showcase the advantages of the proposed MPC approach on the example of a ball-balancing robot with a robotic manipulator and validate our controller in hardware experiments for tasks such as end-effector pose tracking and door opening

    Local Communication Protocols for Learning Complex Swarm Behaviors with Deep Reinforcement Learning

    Full text link
    Swarm systems constitute a challenging problem for reinforcement learning (RL) as the algorithm needs to learn decentralized control policies that can cope with limited local sensing and communication abilities of the agents. While it is often difficult to directly define the behavior of the agents, simple communication protocols can be defined more easily using prior knowledge about the given task. In this paper, we propose a number of simple communication protocols that can be exploited by deep reinforcement learning to find decentralized control policies in a multi-robot swarm environment. The protocols are based on histograms that encode the local neighborhood relations of the agents and can also transmit task-specific information, such as the shortest distance and direction to a desired target. In our framework, we use an adaptation of Trust Region Policy Optimization to learn complex collaborative tasks, such as formation building and building a communication link. We evaluate our findings in a simulated 2D-physics environment, and compare the implications of different communication protocols.Comment: 13 pages, 4 figures, version 2, accepted at ANTS 201
    • …
    corecore