3 research outputs found
Graph Neural Networks for Learning Robot Team Coordination
This paper shows how Graph Neural Networks can be used for learning
distributed coordination mechanisms in connected teams of robots. We capture
the relational aspect of robot coordination by modeling the robot team as a
graph, where each robot is a node, and edges represent communication links.
During training, robots learn how to pass messages and update internal states,
so that a target behavior is reached. As a proxy for more complex problems,
this short paper considers the problem where each robot must locally estimate
the algebraic connectivity of the team's network topology.Comment: Presented at the Federated AI for Robotics Workshop,
IJCAI-ECAI/ICML/AAMAS 201
A Dataset Schema for Cooperative Learning from Demonstration in Multi-robots Systems
Multi-Agent Systems (MASs) have been used to solve complex problems that
demand intelligent agents working together to reach the desired goals. These
Agents should effectively synchronize their individual behaviors so that they
can act as a team in a coordinated manner to achieve the common goal of the
whole system. One of the main issues in MASs is the agents' coordination, being
common domain experts observing MASs execution disapprove agents' decisions.
Even if the MAS was designed using the best methods and tools for agents'
coordination, this difference of decisions between experts and MAS is
confirmed. Therefore, this paper proposes a new dataset schema to support
learning the coordinated behavior in MASs from demonstration. The results of
the proposed solution are validated in a Multi-Robot System (MRS) organizing a
collection of new cooperative plans recommendations from the demonstration by
domain experts.Comment: This is a pre-print of an article published in the Journal of
Intelligent & Robotic Systems. The final authenticated version will be
available online at: https://doi. org/10.1007/s10846-019-01123-
A Survey of Learning in Multiagent Environments: Dealing with Non-Stationarity
The key challenge in multiagent learning is learning a best response to the
behaviour of other agents, which may be non-stationary: if the other agents
adapt their strategy as well, the learning target moves. Disparate streams of
research have approached non-stationarity from several angles, which make a
variety of implicit assumptions that make it hard to keep an overview of the
state of the art and to validate the innovation and significance of new works.
This survey presents a coherent overview of work that addresses
opponent-induced non-stationarity with tools from game theory, reinforcement
learning and multi-armed bandits. Further, we reflect on the principle
approaches how algorithms model and cope with this non-stationarity, arriving
at a new framework and five categories (in increasing order of sophistication):
ignore, forget, respond to target models, learn models, and theory of mind. A
wide range of state-of-the-art algorithms is classified into a taxonomy, using
these categories and key characteristics of the environment (e.g.,
observability) and adaptation behaviour of the opponents (e.g., smooth,
abrupt). To clarify even further we present illustrative variations of one
domain, contrasting the strengths and limitations of each category. Finally, we
discuss in which environments the different approaches yield most merit, and
point to promising avenues of future research.Comment: 64 pages, 7 figures. Under review since November 201