15 research outputs found
Constructing a unifying theory of dynamic programming DCOP algorithms via the generalized distributive law
In this paper we propose a novel message-passing algorithm, the so-called Action-GDL, as an extension to the generalized distributive law (GDL) to ef¿ciently solve DCOPs. Action-GDL provides a unifying perspective of several dynamic programming DCOP algorithms that are based on GDL, such as DPOP and DCPOP algorithms. We empirically show how Action-GDL using a novel distributed post-processing heuristic can outperform DCPOP, and by extension DPOP, even when the latter uses the best arrangement provided by multiple state-of-the-art heuristics.Work funded by IEA (TIN2006-15662-C02-01), AT (CONSOLIDER CSD2007-0022, INGENIO 2010) and EVE (TIN2009-14702-C02-01 and 02). Vinyals is supported by the Spanish Ministry of Education (FPU grant AP2006-04636)Peer Reviewe
Implementing MAS agreement processes based on consensus networks
[EN] Consensus is a negotiation process where agents need to agree upon certain quantities of interest. The theoretical framework for solving consensus problems in dynamic networks of agents was formally introduced by Olfati-Saber and Murray, and is based on algebraic graph theory, matrix theory and control theory. Consensus problems are usually simulated using mathematical frameworks. However, implementation using multi-agent system platforms is a very difficult task due to problems such as synchronization, distributed finalization, and monitorization among others. The aim of this paper is to propose a protocol for the consensus agreement process in MAS in order to check the correctness of the algorithm and validate the protocol. © Springer International Publishing Switzerland 2013.This work is supported by ww and PROMETEO/2008/051 projects of the Spanish government, CONSOLIDER-INGENIO 2010 under grant CSD2007-00022, TIN2012-36586-C03-01 and PAID-06-11-2084.Palomares Chust, A.; Carrascosa Casamayor, C.; Rebollo Pedruelo, M.; Gómez, Y. (2013). Implementing MAS agreement processes based on consensus networks. Distributed Computing and Artificial Intelligence. 217:553-560. https://doi.org/10.1007/978-3-319-00551-5_66S553560217Argente, E.: et al: An Abstract Architecture for Virtual Organizations: The THOMAS approach. Knowledge and Information Systems 29(2), 379–403 (2011)Búrdalo, L.: et al: TRAMMAS: A tracing model for multiagent systems. Eng. Appl. Artif. Intel. 24(7), 1110–1119 (2011)Fogués, R.L., et al.: Towards Dynamic Agent Interaction Support in Open Multiagent Systems. In: Proc. of the 13th CCIA, vol. 220, pp. 89–98. IOS Press (2010)Luck, M., et al.: Agent technology: Computing as interaction (a roadmap for agent based computing). Eng. Appl. Artif. Intel. (2005)Mailler, R., Lesser, V.: Solving distributed constraint optimization problems using cooperative mediation. In: AAMAS 2004, pp. 438–445 (2004)Olfati-Saber, R., Fax, J.A., Murray, R.M.: Consensus and cooperation in networked multi-agent systems. Proceedings of the IEEE 95(1), 215–233 (2007)Pujol-Gonzalez, M.: Multi-agent coordination: Dcops and beyond. In: Proc. of IJCAI, pp. 2838–2839 (2011)Such, J.: et al: Magentix2: A privacy-enhancing agent platform. Eng. Appl. Artif. Intel. 26(1), 96–109 (2013)Vinyals, M., et al.: Constructing a unifying theory of dynamic programming dcop algorithms via the generalized distributive law. Autonomous Agents and Multi-Agent Systems 22, 439–464 (2011
Embedding Preference Elicitation Within the Search for DCOP Solutions
The Distributed Constraint Optimization Problem(DCOP)formulation is a powerful tool to model cooperative multi-agent problems, especially when they are sparsely constrained with one another. A key assumption in this model is that all constraints are fully specified or known a priori, which may not hold in applications where constraints encode preferences of human users. In this thesis, we extend the model to Incomplete DCOPs (I-DCOPs), where some constraints can be partially specified. User preferences for these partially-specified constraints can be elicited during the execution of I-DCOP algorithms, but they incur some elicitation costs. Additionally, we propose two parameterized heuristics that can be used in conjunction with Synchronous Branch-and-Bound to solve I-DCOPs. These heuristics allow users to trade-off solution quality for faster runtimes and a smaller number of elicitations. They also provide theoretical quality guarantees for problems where elicitations are free. Our model and heuristics thus extend the state of the art in distributed constraint reasoning to better model and solve distributed agent-based applications with user preferences
Recommended from our members
Application of Techniques for MAP Estimation to Distributed Constraint Optimization Problem
The problem of efficiently finding near-optimal decisions in multi-agent systems has become increasingly important because of the growing number of multi-agent applications with large numbers of agents operating in real-world environments. In these systems, agents are often subject to tight resource constraints and agents have only local views. When agents have non-global constraints, each of which is independent, the problem can be formalized as a distributed constraint optimization problem (DCOP). The DCOP is closely associated with the problem of inference on graphical models. Many approaches from inference literature have been adopted to solve DCOPs. We focus on the Max-Sum algorithm and the Action-GDL algorithm that are DCOP variants of the popular inference algorithm called the Max-Product algorithm and the Belief Propagation algorithm respectively. The Max-Sum algorithm and the Action-GDL algorithm are well-suited for multi-agent systems because it is distributed by nature and requires less communication than most DCOP algorithms. However, the resource requirements of these algorithms are still high for some multi-agent domains and various aspects of the algorithms have not been well studied for use in general multi-agent settings.
This thesis is concerned with a variety of issues of applying the Max-Sum algorithms and the Action-GDL algorithm to general multi-agent settings. We develop a hybrid algorithm of ADOPT and Action-GDL in order to overcome the communication complexity of DCOPs. Secondly, we extend the Max-Sum algorithm to operate more efficiently in more general multi-agent settings in which computational complexity is high. We provide an algorithm that has a lower expected computational complexity for DCOPs even with n-ary constraints. Finally, In most DCOP literature, a one-to-one mapping between a variable and an agent is assumed. However, in real applications, many-to-one mappings are prevalent and can also be beneficial in terms of communication and hardware cost in situations where agents are acting as independent computing units. We consider how to exploit such mapping in order to increase efficiency
Distributed Gibbs: A memory-bounded sampling-based DCOP algorithm
National Research Foundation (NRF) Singapore under International Research Centres in Singapore Funding Initiativ
CUBE: A CUDA approach for Bucket Elimination on GPUs
We consider Bucket Elimination (BE), a popular algorithmic framework to solve Constraint Optimisation Problems (COPs). We focus on the parallelisation of the most computationally intensive operations of BE, i.e., join sum and maximisation, which are key ingredients in several close variants of the BE framework (including Belief Propagation on Junction Trees and Distributed COP techniques such as ActionGDL and DPOP). In particular, we propose CUBE, a highly-parallel GPU implementation of such operations, which adopts an efficient memory layout allowing all threads to independently locate their input and output addresses in memory, hence achieving a high computational throughput. We compare CUBE with the most recent GPU implementation of BE. Our results show that CUBE achieves significant speed-ups (up to two orders of magnitude) w.r.t. the counterpart approach, showing a dramatic decrease of the runtime w.r.t. the serial version (i.e., up to 652x faster). More important, such speed-ups increase when the complexity of the problem grows, showing that CUBE correctly exploits the additional degree of parallelism inherent in the problem
Distributed Constraint Optimization:Privacy Guarantees and Stochastic Uncertainty
Distributed Constraint Satisfaction (DisCSP) and Distributed Constraint Optimization (DCOP) are formal frameworks that can be used to model a variety of problems in which multiple decision-makers cooperate towards a common goal: from computing an equilibrium of a game, to vehicle routing problems, to combinatorial auctions. In this thesis, we independently address two important issues in such multi-agent problems: 1) how to provide strong guarantees on the protection of the privacy of the participants, and 2) how to anticipate future, uncontrollable events. On the privacy front, our contributions depart from previous work in two ways. First, we consider not only constraint privacy (the agents' private costs) and decision privacy (keeping the complete solution secret), but also two other types of privacy that have been largely overlooked in the literature: agent privacy, which has to do with protecting the identities of the participants, and topology privacy, which covers information about the agents' co-dependencies. Second, while previous work focused mainly on quantitatively measuring and reducing privacy loss, our algorithms provide stronger, qualitative guarantees on what information will remain secret. Our experiments show that it is possible to provide such privacy guarantees, while still scaling to much larger problems than the previous state of the art. When it comes to reasoning under uncertainty, we propose an extension to the DCOP framework, called DCOP under Stochastic Uncertainty (StochDCOP), which includes uncontrollable, random variables with known probability distributions that model uncertain, future events. The problem becomes one of making "optimal" offline decisions, before the true values of the random variables can be observed. We consider three possible concepts of optimality: minimizing the expected cost, minimizing the worst-case cost, or maximizing the probability of a-posteriori optimality. We propose a new family of StochDCOP algorithms, exploring the tradeoffs between solution quality, computational and message complexity, and privacy. In particular, we show how discovering and reasoning about co-dependencies on common random variables can yield higher-quality solutions