6,075 research outputs found
Decentralized MPC based Obstacle Avoidance for Multi-Robot Target Tracking Scenarios
In this work, we consider the problem of decentralized multi-robot target
tracking and obstacle avoidance in dynamic environments. Each robot executes a
local motion planning algorithm which is based on model predictive control
(MPC). The planner is designed as a quadratic program, subject to constraints
on robot dynamics and obstacle avoidance. Repulsive potential field functions
are employed to avoid obstacles. The novelty of our approach lies in embedding
these non-linear potential field functions as constraints within a convex
optimization framework. Our method convexifies non-convex constraints and
dependencies, by replacing them as pre-computed external input forces in robot
dynamics. The proposed algorithm additionally incorporates different methods to
avoid field local minima problems associated with using potential field
functions in planning. The motion planner does not enforce predefined
trajectories or any formation geometry on the robots and is a comprehensive
solution for cooperative obstacle avoidance in the context of multi-robot
target tracking. We perform simulation studies in different environmental
scenarios to showcase the convergence and efficacy of the proposed algorithm.
Video of simulation studies: \url{https://youtu.be/umkdm82Tt0M
Towards Optimally Decentralized Multi-Robot Collision Avoidance via Deep Reinforcement Learning
Developing a safe and efficient collision avoidance policy for multiple
robots is challenging in the decentralized scenarios where each robot generate
its paths without observing other robots' states and intents. While other
distributed multi-robot collision avoidance systems exist, they often require
extracting agent-level features to plan a local collision-free action, which
can be computationally prohibitive and not robust. More importantly, in
practice the performance of these methods are much lower than their centralized
counterparts.
We present a decentralized sensor-level collision avoidance policy for
multi-robot systems, which directly maps raw sensor measurements to an agent's
steering commands in terms of movement velocity. As a first step toward
reducing the performance gap between decentralized and centralized methods, we
present a multi-scenario multi-stage training framework to find an optimal
policy which is trained over a large number of robots on rich, complex
environments simultaneously using a policy gradient based reinforcement
learning algorithm. We validate the learned sensor-level collision avoidance
policy in a variety of simulated scenarios with thorough performance
evaluations and show that the final learned policy is able to find time
efficient, collision-free paths for a large-scale robot system. We also
demonstrate that the learned policy can be well generalized to new scenarios
that do not appear in the entire training period, including navigating a
heterogeneous group of robots and a large-scale scenario with 100 robots.
Videos are available at https://sites.google.com/view/drlmac
- …