16,345 research outputs found

    Control Architecture for Cooperative Mobile Robots using Multi-Agent based Coordination Approach

    Get PDF
    National audienceThis paper is about a Multi-Agent based solution to control and coordinate team-working mobile robots moving in unstructured environments. Two main contributions are considered in our approach. The rst contribution of this paper is about the Multi-Agents System to Control and Coordinate teAmworking Robots (MAS2CAR) architecture, a new architecture to control a group of coordinated autonomous robots in unstructured environments. MAS2CAR covers three main layers: (i) the Physical Layer (ii) the Control Layer and (iii) the Coordination Layer. The second contribution of this paper is about the multi-agent system (MAS) organisational models aiming to solve the key cooperation issues in the coordination layer, the software components designed based on Utopia a MAS framework which automatically build software agents, thanks to a multi-agent based organisational model called MoiseInst . We provide simulation results that exhibit robotics cooperative behavior related to our scenario, such as multi-robots navigation in presence of obstacles (including trajectory planning, and reactive aspects) via a hybrid control

    Danger-aware Adaptive Composition of DRL Agents for Self-navigation

    Full text link
    Self-navigation, referred as the capability of automatically reaching the goal while avoiding collisions with obstacles, is a fundamental skill required for mobile robots. Recently, deep reinforcement learning (DRL) has shown great potential in the development of robot navigation algorithms. However, it is still difficult to train the robot to learn goal-reaching and obstacle-avoidance skills simultaneously. On the other hand, although many DRL-based obstacle-avoidance algorithms are proposed, few of them are reused for more complex navigation tasks. In this paper, a novel danger-aware adaptive composition (DAAC) framework is proposed to combine two individually DRL-trained agents, obstacle-avoidance and goal-reaching, to construct a navigation agent without any redesigning and retraining. The key to this adaptive composition approach is that the value function outputted by the obstacle-avoidance agent serves as an indicator for evaluating the risk level of the current situation, which in turn determines the contribution of these two agents for the next move. Simulation and real-world testing results show that the composed Navigation network can control the robot to accomplish difficult navigation tasks, e.g., reaching a series of successive goals in an unknown and complex environment safely and quickly.Comment: 7 pages, 9 figure

    Towards Optimally Decentralized Multi-Robot Collision Avoidance via Deep Reinforcement Learning

    Full text link
    Developing a safe and efficient collision avoidance policy for multiple robots is challenging in the decentralized scenarios where each robot generate its paths without observing other robots' states and intents. While other distributed multi-robot collision avoidance systems exist, they often require extracting agent-level features to plan a local collision-free action, which can be computationally prohibitive and not robust. More importantly, in practice the performance of these methods are much lower than their centralized counterparts. We present a decentralized sensor-level collision avoidance policy for multi-robot systems, which directly maps raw sensor measurements to an agent's steering commands in terms of movement velocity. As a first step toward reducing the performance gap between decentralized and centralized methods, we present a multi-scenario multi-stage training framework to find an optimal policy which is trained over a large number of robots on rich, complex environments simultaneously using a policy gradient based reinforcement learning algorithm. We validate the learned sensor-level collision avoidance policy in a variety of simulated scenarios with thorough performance evaluations and show that the final learned policy is able to find time efficient, collision-free paths for a large-scale robot system. We also demonstrate that the learned policy can be well generalized to new scenarios that do not appear in the entire training period, including navigating a heterogeneous group of robots and a large-scale scenario with 100 robots. Videos are available at https://sites.google.com/view/drlmac

    Multi-robot team formation control in the GUARDIANS project

    Get PDF
    Purpose The GUARDIANS multi-robot team is to be deployed in a large warehouse in smoke. The team is to assist firefighters search the warehouse in the event or danger of a fire. The large dimensions of the environment together with development of smoke which drastically reduces visibility, represent major challenges for search and rescue operations. The GUARDIANS robots guide and accompany the firefighters on site whilst indicating possible obstacles and the locations of danger and maintaining communications links. Design/methodology/approach In order to fulfill the aforementioned tasks the robots need to exhibit certain behaviours. Among the basic behaviours are capabilities to stay together as a group, that is, generate a formation and navigate while keeping this formation. The control model used to generate these behaviours is based on the so-called social potential field framework, which we adapt to the specific tasks required for the GUARDIANS scenario. All tasks can be achieved without central control, and some of the behaviours can be performed without explicit communication between the robots. Findings The GUARDIANS environment requires flexible formations of the robot team: the formation has to adapt itself to the circumstances. Thus the application has forced us to redefine the concept of a formation. Using the graph-theoretic terminology, we can say that a formation may be stretched out as a path or be compact as a star or wheel. We have implemented the developed behaviours in simulation environments as well as on real ERA-MOBI robots commonly referred to as Erratics. We discuss advantages and shortcomings of our model, based on the simulations as well as on the implementation with a team of Erratics.</p

    Experiments in cooperative human multi-robot navigation

    Get PDF
    In this paper, we consider the problem of a group of autonomous mobile robots and a human moving coordinately in a real-world implementation. The group moves throughout a dynamic and unstructured environment. The key problem to be solved is the inclusion of a human in a real multi-robot system and consequently the multiple robot motion coordination. We present a set of performance metrics (system efficiency and percentage of time in formation) and a novel flexible formation definition whereby a formation control strategy both in simulation and in real-world experiments of a human multi-robot system is presented. The formation control proposed is stable and effective by means of its uniform dispersion, cohesion and flexibility
    corecore