2,389 research outputs found
Bounded Decentralised Coordination over Multiple Objectives
We propose the bounded multi-objective max-sum algorithm (B-MOMS), the first decentralised coordination algorithm for multi-objective optimisation problems. B-MOMS extends the max-sum message-passing algorithm for decentralised coordination to compute bounded approximate solutions to multi-objective decentralised constraint optimisation problems (MO-DCOPs). Specifically, we prove the optimality of B-MOMS in acyclic constraint graphs, and derive problem dependent bounds on its approximation ratio when these graphs contain cycles. Furthermore, we empirically evaluate its performance on a multi-objective extension of the canonical graph colouring problem. In so doing, we demonstrate that, for the settings we consider, the approximation ratio never exceeds 2, and is typically less than 1.5 for less-constrained graphs. Moreover, the runtime required by B-MOMS on the problem instances we considered never exceeds 30 minutes, even for maximally constrained graphs with agents. Thus, B-MOMS brings the problem of multi-objective optimisation well within the boundaries of the limited capabilities of embedded agents
Coordinated Multi-Agent Imitation Learning
We study the problem of imitation learning from demonstrations of multiple
coordinating agents. One key challenge in this setting is that learning a good
model of coordination can be difficult, since coordination is often implicit in
the demonstrations and must be inferred as a latent variable. We propose a
joint approach that simultaneously learns a latent coordination model along
with the individual policies. In particular, our method integrates unsupervised
structure learning with conventional imitation learning. We illustrate the
power of our approach on a difficult problem of learning multiple policies for
fine-grained behavior modeling in team sports, where different players occupy
different roles in the coordinated team strategy. We show that having a
coordination model to infer the roles of players yields substantially improved
imitation loss compared to conventional baselines.Comment: International Conference on Machine Learning 201
Exploiting Anonymity in Approximate Linear Programming: Scaling to Large Multiagent MDPs (Extended Version)
Many exact and approximate solution methods for Markov Decision Processes
(MDPs) attempt to exploit structure in the problem and are based on
factorization of the value function. Especially multiagent settings, however,
are known to suffer from an exponential increase in value component sizes as
interactions become denser, meaning that approximation architectures are
restricted in the problem sizes and types they can handle. We present an
approach to mitigate this limitation for certain types of multiagent systems,
exploiting a property that can be thought of as "anonymous influence" in the
factored MDP. Anonymous influence summarizes joint variable effects efficiently
whenever the explicit representation of variable identity in the problem can be
avoided. We show how representational benefits from anonymity translate into
computational efficiencies, both for general variable elimination in a factor
graph but in particular also for the approximate linear programming solution to
factored MDPs. The latter allows to scale linear programming to factored MDPs
that were previously unsolvable. Our results are shown for the control of a
stochastic disease process over a densely connected graph with 50 nodes and 25
agents.Comment: Extended version of AAAI 2016 pape
A Hybrid Three Layer Architecture for Fire Agent Management in Rescue Simulation Environment
This paper presents a new architecture called FAIS for imple- menting
intelligent agents cooperating in a special Multi Agent environ- ment, namely
the RoboCup Rescue Simulation System. This is a layered architecture which is
customized for solving fire extinguishing problem. Structural decision making
algorithms are combined with heuristic ones in this model, so it's a hybrid
architecture
Distributed Consensus of Linear Multi-Agent Systems with Switching Directed Topologies
This paper addresses the distributed consensus problem for a linear
multi-agent system with switching directed communication topologies. By
appropriately introducing a linear transformation, the consensus problem is
equivalently converted to a stabilization problem for a class of switched
linear systems. Some sufficient consensus conditions are then derived by using
tools from the matrix theory and stability analysis of switched systems. It is
proved that consensus in such a multi-agent system can be ensured if each agent
is stabilizable and each possible directed topology contains a directed
spanning tree. Finally, a numerical simulation is given for illustration.Comment: The paper will be presented at the 2014 Australian Control Conference
(AUCC 2014), Canberra, Australi
- …