7,010 research outputs found
Deep Reinforcement Learning for Swarm Systems
Recently, deep reinforcement learning (RL) methods have been applied
successfully to multi-agent scenarios. Typically, these methods rely on a
concatenation of agent states to represent the information content required for
decentralized decision making. However, concatenation scales poorly to swarm
systems with a large number of homogeneous agents as it does not exploit the
fundamental properties inherent to these systems: (i) the agents in the swarm
are interchangeable and (ii) the exact number of agents in the swarm is
irrelevant. Therefore, we propose a new state representation for deep
multi-agent RL based on mean embeddings of distributions. We treat the agents
as samples of a distribution and use the empirical mean embedding as input for
a decentralized policy. We define different feature spaces of the mean
embedding using histograms, radial basis functions and a neural network learned
end-to-end. We evaluate the representation on two well known problems from the
swarm literature (rendezvous and pursuit evasion), in a globally and locally
observable setup. For the local setup we furthermore introduce simple
communication protocols. Of all approaches, the mean embedding representation
using neural network features enables the richest information exchange between
neighboring agents facilitating the development of more complex collective
strategies.Comment: 31 pages, 12 figures, version 3 (published in JMLR Volume 20
New Game-Theoretic Convolutional Neural Network Applied for the Multi-Pursuer Multi-Evader Game
Pursuit-Evasion Game (PEG) can be defined as a set of agents known as pursuers, which cooperate with the aim forming dynamic coalitions to capture dynamic evader agents, while the evaders try to avoid this capture by moving in the environment according to specific velocities. The factor of capturing time was treated by various studies before, but remain the powerful tools used to satisfy this factor object of research. To improve the capturing time factor we proposed in this work a novel online decentralized coalition formation algorithm equipped with Convolutional Neural Network (CNN) and based on the Iterated Elimination of Dominated Strategies (IEDS). The coalition is formed such that the pursuer should learn at each iteration the approximator formation achieving the capture in the shortest time. The pursuer’s learning process depends on the features extracted by CNN at each iteration. The proposed supervised technique is compared through simulation, with the IEDS algorithm, AGR algorithm. Simulation results show that the proposed learning technique outperform the IEDS algorithm and the AGR algorithm with respect to the learning time which represents an important factor in a chasing game
Intelligent Escape of Robotic Systems: A Survey of Methodologies, Applications, and Challenges
Intelligent escape is an interdisciplinary field that employs artificial
intelligence (AI) techniques to enable robots with the capacity to
intelligently react to potential dangers in dynamic, intricate, and
unpredictable scenarios. As the emphasis on safety becomes increasingly
paramount and advancements in robotic technologies continue to advance, a wide
range of intelligent escape methodologies has been developed in recent years.
This paper presents a comprehensive survey of state-of-the-art research work on
intelligent escape of robotic systems. Four main methods of intelligent escape
are reviewed, including planning-based methodologies, partitioning-based
methodologies, learning-based methodologies, and bio-inspired methodologies.
The strengths and limitations of existing methods are summarized. In addition,
potential applications of intelligent escape are discussed in various domains,
such as search and rescue, evacuation, military security, and healthcare. In an
effort to develop new approaches to intelligent escape, this survey identifies
current research challenges and provides insights into future research trends
in intelligent escape.Comment: This paper is accepted by Journal of Intelligent and Robotic System
Fairness for Cooperative Multi-Agent Learning with Equivariant Policies
We study fairness through the lens of cooperative multi-agent learning. Our
work is motivated by empirical evidence that naive maximization of team reward
yields unfair outcomes for individual team members. To address fairness in
multi-agent contexts, we introduce team fairness, a group-based fairness
measure for multi-agent learning. We then incorporate team fairness into policy
optimization -- introducing Fairness through Equivariance (Fair-E), a novel
learning strategy that achieves provably fair reward distributions. We then
introduce Fairness through Equivariance Regularization (Fair-ER) as a
soft-constraint version of Fair-E and show that Fair-ER reaches higher levels
of utility than Fair-E and fairer outcomes than policies with no equivariance.
Finally, we investigate the fairness-utility trade-off in multi-agent settings.Comment: 15 pages, 4 figure
- …