340 research outputs found
Multi-agent Motion Planning for Dense and Dynamic Environments via Deep Reinforcement Learning
This paper introduces a hybrid algorithm of deep reinforcement learning (RL)
and Force-based motion planning (FMP) to solve distributed motion planning
problem in dense and dynamic environments. Individually, RL and FMP algorithms
each have their own limitations. FMP is not able to produce time-optimal paths
and existing RL solutions are not able to produce collision-free paths in dense
environments. Therefore, we first tried improving the performance of recent RL
approaches by introducing a new reward function that not only eliminates the
requirement of a pre supervised learning (SL) step but also decreases the
chance of collision in crowded environments. That improved things, but there
were still a lot of failure cases. So, we developed a hybrid approach to
leverage the simpler FMP approach in stuck, simple and high-risk cases, and
continue using RL for normal cases in which FMP can't produce optimal path.
Also, we extend GA3C-CADRL algorithm to 3D environment. Simulation results show
that the proposed algorithm outperforms both deep RL and FMP algorithms and
produces up to 50% more successful scenarios than deep RL and up to 75% less
extra time to reach goal than FMP.Comment: IEEE Robotics and Automation Letters (2020
A Survey on Passing-through Control of Multi-Robot Systems in Cluttered Environments
This survey presents a comprehensive review of various methods and algorithms
related to passing-through control of multi-robot systems in cluttered
environments. Numerous studies have investigated this area, and we identify
several avenues for enhancing existing methods. This survey describes some
models of robots and commonly considered control objectives, followed by an
in-depth analysis of four types of algorithms that can be employed for
passing-through control: leader-follower formation control, multi-robot
trajectory planning, control-based methods, and virtual tube planning and
control. Furthermore, we conduct a comparative analysis of these techniques and
provide some subjective and general evaluations.Comment: 18 pages, 19 figure
Decentralized Triangular Guidance Algorithms for Formations of UAVs
This paper deals with the design of a guidance control system for a swarm of unmanned aerial systems flying at a given altitude, addressing flight formation requirements that can be formulated constraining the swarm to be on the nodes of a triangular mesh. Three decentralized guidance algorithms are presented. A classical fixed leader–follower scheme is compared with two alternative
schemes: the former is based on the self-identification of one or more time-varying leaders; the latter is an algorithm without leaders. Several operational scenarios have been simulated involving swarms with obstacles and an increasing number of aircraft in order to prove the effectiveness of the proposed guidance schem
A Survey on Aerial Swarm Robotics
The use of aerial swarms to solve real-world problems has been increasing steadily, accompanied by falling prices and improving performance of communication, sensing, and processing hardware. The commoditization of hardware has reduced unit costs, thereby lowering the barriers to entry to the field of aerial swarm robotics. A key enabling technology for swarms is the family of algorithms that allow the individual members of the swarm to communicate and allocate tasks amongst themselves, plan their trajectories, and coordinate their flight in such a way that the overall objectives of the swarm are achieved efficiently. These algorithms, often organized in a hierarchical fashion, endow the swarm with autonomy at every level, and the role of a human operator can be reduced, in principle, to interactions at a higher level without direct intervention. This technology depends on the clever and innovative application of theoretical tools from control and estimation. This paper reviews the state of the art of these theoretical tools, specifically focusing on how they have been developed for, and applied to, aerial swarms. Aerial swarms differ from swarms of ground-based vehicles in two respects: they operate in a three-dimensional space and the dynamics of individual vehicles adds an extra layer of complexity. We review dynamic modeling and conditions for stability and controllability that are essential in order to achieve cooperative flight and distributed sensing. The main sections of this paper focus on major results covering trajectory generation, task allocation, adversarial control, distributed sensing, monitoring, and mapping. Wherever possible, we indicate how the physics and subsystem technologies of aerial robots are brought to bear on these individual areas
Adaptive and learning-based formation control of swarm robots
Autonomous aerial and wheeled mobile robots play a major role in tasks such as search and rescue, transportation, monitoring, and inspection. However, these operations are faced with a few open challenges including robust autonomy, and adaptive coordination based on the environment and operating conditions, particularly in swarm robots with limited communication and perception capabilities. Furthermore, the computational complexity increases exponentially with the number of robots in the swarm. This thesis examines two different aspects of the formation control problem. On the one hand, we investigate how formation could be performed by swarm robots with limited communication and perception (e.g., Crazyflie nano quadrotor). On the other hand, we explore human-swarm interaction (HSI) and different shared-control mechanisms between human and swarm robots (e.g., BristleBot) for artistic creation. In particular, we combine bio-inspired (i.e., flocking, foraging) techniques with learning-based control strategies (using artificial neural networks) for adaptive control of multi- robots. We first review how learning-based control and networked dynamical systems can be used to assign distributed and decentralized policies to individual robots such that the desired formation emerges from their collective behavior. We proceed by presenting a novel flocking control for UAV swarm using deep reinforcement learning. We formulate the flocking formation problem as a partially observable Markov decision process (POMDP), and consider a leader-follower configuration, where consensus among all UAVs is used to train a shared control policy, and each UAV performs actions based on the local information it collects. In addition, to avoid collision among UAVs and guarantee flocking and navigation, a reward function is added with the global flocking maintenance, mutual reward, and a collision penalty. We adapt deep deterministic policy gradient (DDPG) with centralized training and decentralized execution to obtain the flocking control policy using actor-critic networks and a global state space matrix. In the context of swarm robotics in arts, we investigate how the formation paradigm can serve as an interaction modality for artists to aesthetically utilize swarms. In particular, we explore particle swarm optimization (PSO) and random walk to control the communication between a team of robots with swarming behavior for musical creation
- …