657 research outputs found

    Coordinating decentralized learning and conflict resolution across agent boundaries

    Get PDF
    It is crucial for embedded systems to adapt to the dynamics of open environments. This adaptation process becomes especially challenging in the context of multiagent systems because of scalability, partial information accessibility and complex interaction of agents. It is a challenge for agents to learn good policies, when they need to plan and coordinate in uncertain, dynamic environments, especially when they have large state spaces. It is also critical for agents operating in a multiagent system (MAS) to resolve conflicts among the learned policies of different agents, since such conflicts may have detrimental influence on the overall performance. The focus of this research is to use a reinforcement learning based local optimization algorithm within each agent to learn multiagent policies in a decentralized fashion. These policies will allow each agent to adapt to changes in environmental conditions while reorganizing the underlying multiagent network when needed. The research takes an adaptive approach to resolving conflicts that can arise between locally optimal agent policies. First an algorithm that uses heuristic rules to locally resolve simple conflicts is presented. When the environment is more dynamic and uncertain, a mediator-based mechanism to resolve more complicated conflicts and selectively expand the agents' state space during the learning process is harnessed. For scenarios where mediator-based mechanisms with partially global views are ineffective, a more rigorous approach for global conflict resolution that synthesizes multiagent reinforcement learning (MARL) and distributed constraint optimization (DCOP) is developed. These mechanisms are evaluated in the context of a multiagent tornado tracking application called NetRads. Empirical results show that these mechanisms significantly improve the performance of the tornado tracking network for a variety of weather scenarios. The major contributions of this work are: a state of the art decentralized learning approach that supports agent interactions and reorganizes the underlying network when needed; the use of abstract classes of scenarios/states/actions that efficiently manages the exploration of the search space; novel conflict resolution algorithms of increasing complexity that use heuristic rules, sophisticated automated negotiation mechanisms and distributed constraint optimization methods respectively; and finally, a rigorous study of the interplay between two popular theories used to solve multiagent problems, namely decentralized Markov decision processes and distributed constraint optimization

    Multi-platform coordination and resource management in command and control

    Get PDF
    Depuis plusieurs années, nous constatons l'augmentation de l'utilisation des techniques d'agents et multiagent pour assister l'humain dans ses tâches. Ce travail de maîtrise se situe dans la même voie. Précisément, nous proposons d'utiliser les techniques multiagent de planification et de coordination pour la gestion de ressources dans les systèmes de commande et contrôle (C2) temps réel. Le problème particulier que nous avons étudié est la conception d'un système d'aide à la décision pour les opérations anti-aérienne sur les frégates canadiennes. Dans le cas où plusieurs frégates doivent se défendre contre des menaces, la coordination est un problème d'importance capitale. L'utilisation de mécanismes de coordination efficaces permet d'éviter les actions conflictuelles et la redondance dans les engagements. Dans ce mémoire, nous présentons quatre mécanismes de coordination basés sur le partage de tâche. Trois sont basés sur les communications : la coordination centrale, le Contract Net, la coordination similaire à celle proposée par Brown; tandis que la défense de zone est basée sur les lois sociales. Nous exposons enfin les résultats auxquels nous sommes arrivés en simulant ces différents mécanismes.The use of agent and multiagent techniques to assist humans in their daily routines has been increasing for many years, notably in Command and Control (C2) systems. This thesis is is situated in this domain. Precisely, we propose to use multiagent planning and coordination techniques for resource management in real-time \acs{C2} systems. The particular problem we studied is the design of a decision-support for anti-air warfare on Canadian frigates. In the case of several frigates defending against incoming threats, multiagent coordination is a complex problem of capital importance. Better coordination mechanisms are important to avoid redundancy in engagements and inefficient defence caused by conflicting actions. In this thesis, we present four different coordination mechanisms based on task sharing. Three of these mechanisms are based on communications: central coordination, Contract Net coordination and Brown coordination, while the zone defence coordination is based on social laws. Finally, we expose the results obtained while simulating these various mechanisms

    Emergence and resilience in multi-agent reinforcement learning

    Get PDF
    Our world represents an enormous multi-agent system (MAS), consisting of a plethora of agents that make decisions under uncertainty to achieve certain goals. The interaction of agents constantly affects our world in various ways, leading to the emergence of interesting phenomena like life forms and civilizations that can last for many years while withstanding various kinds of disturbances. Building artificial MAS that are able to adapt and survive similarly to natural MAS is a major goal in artificial intelligence as a wide range of potential real-world applications like autonomous driving, multi-robot warehouses, and cyber-physical production systems can be straightforwardly modeled as MAS. Multi-agent reinforcement learning (MARL) is a promising approach to build such systems which has achieved remarkable progress in recent years. However, state-of-the-art MARL commonly assumes very idealized conditions to optimize performance in best-case scenarios while neglecting further aspects that are relevant to the real world. In this thesis, we address emergence and resilience in MARL which are important aspects to build artificial MAS that adapt and survive as effectively as natural MAS do. We first focus on emergent cooperation from local interaction of self-interested agents and introduce a peer incentivization approach based on mutual acknowledgments. We then propose to exploit emergent phenomena to further improve coordination in large cooperative MAS via decentralized planning or hierarchical value function factorization. To maintain multi-agent coordination in the presence of partial changes similar to classic distributed systems, we present adversarial methods to improve and evaluate resilience in MARL. Finally, we briefly cover a selection of further topics that are relevant to advance MARL towards real-world applicability.Unsere Welt stellt ein riesiges Multiagentensystem (MAS) dar, welches aus einer Vielzahl von Agenten besteht, die unter Unsicherheit Entscheidungen treffen müssen, um bestimmte Ziele zu erreichen. Die Interaktion der Agenten beeinflusst unsere Welt stets auf unterschiedliche Art und Weise, wodurch interessante emergente Phänomene wie beispielsweise Lebensformen und Zivilisationen entstehen, die über viele Jahre Bestand haben und dabei unterschiedliche Arten von Störungen überwinden können. Die Entwicklung von künstlichen MAS, die ähnlich anpassungs- und überlebensfähig wie natürliche MAS sind, ist eines der Hauptziele in der künstlichen Intelligenz, da viele potentielle Anwendungen wie zum Beispiel das autonome Fahren, die multi-robotergesteuerte Verwaltung von Lagerhallen oder der Betrieb von cyber-phyischen Produktionssystemen, direkt als MAS formuliert werden können. Multi-Agent Reinforcement Learning (MARL) ist ein vielversprechender Ansatz, mit dem in den letzten Jahren bemerkenswerte Fortschritte erzielt wurden, um solche Systeme zu entwickeln. Allerdings geht der Stand der Forschung aktuell von sehr idealisierten Annahmen aus, um die Effektivität ausschließlich für Szenarien im besten Fall zu optimieren. Dabei werden weiterführende Aspekte, die für die echte Welt relevant sind, größtenteils außer Acht gelassen. In dieser Arbeit werden die Aspekte Emergenz und Resilienz in MARL betrachtet, welche wichtig für die Entwicklung von anpassungs- und überlebensfähigen künstlichen MAS sind. Es wird zunächst die Entstehung von emergenter Kooperation durch lokale Interaktion von selbstinteressierten Agenten untersucht. Dazu wird ein Ansatz zur Peer-Incentivierung vorgestellt, welcher auf gegenseitiger Anerkennung basiert. Anschließend werden Ansätze zur Nutzung emergenter Phänomene für die Koordinationsverbesserung in großen kooperativen MAS präsentiert, die dezentrale Planungsverfahren oder hierarchische Faktorisierung von Evaluationsfunktionen nutzen. Zur Aufrechterhaltung der Multiagentenkoordination bei partiellen Veränderungen, ähnlich wie in klassischen verteilten Systemen, werden Methoden des Adversarial Learning vorgestellt, um die Resilienz in MARL zu verbessern und zu evaluieren. Abschließend wird kurz eine Auswahl von weiteren Themen behandelt, die für die Einsatzfähigkeit von MARL in der echten Welt relevant sind

    Strategic Executions of Choreographed Timed Normative Multi-Agent Systems

    Get PDF
    This paper proposes a combined mechanism for coordinating agents in timed normative multi-agent systems. Timing constraints in a multi-agent system make it possible to force action execution to happen before certain time invariants are violated. In such multiagent systems we achieve coordination at two orthogonal levels with respect to states and actions. On the one hand, the behaviour of individual agents is regulated by means of social and organisational inspired concepts like norms and sanctions. On the other hand, the behaviour of sets of agents is restricted according to action-based coordination mechanisms called choreographies. In both cases, the resulting behaviour is constrained by time

    Strategic Structural Reorganization in Multi-agent Systems Inspired by Social Organization Theory

    Get PDF
    Autonomic systems, capable of adaptive behavior, are envisioned as a solution for maintaining large, complex, real-time computing systems that are situated in dynamic and open environments. These systems are subject to uncertainties in their perceptual, computational, and communication loads. As a result, the individual system components find the need to cooperate with each other to acquire more information and accomplish complex tasks. Critical to the effective performance of these systems, is the effectiveness of communication and coordination methods. In many practical applications of distributed and multi-agent systems, the problem of communication and coordination becomes even more complicated because of the geographic disparity of tasks and/or agents that are performing the tasks. Experience with even small systems has shown that lack of an effective communication and coordination strategy leads the system to no-answer, or sub-optimal answer situations. To address this problem, many large-scale systems employ an additional layer of structuring, known as organizational structure, which governs assignment of roles to individual agents, existence of relations between the agents , and any authority structures in between. Applying different organizational structures to the same problem will lead to different performance characteristics. As the system and environment conditions change, it becomes important to reorganize to a more effective organization. Due to the costs associated with reorganization, finding a balance in how often or when a reorganization is performed becomes necessary. In multi-agent systems community, not a lot of attention has been paid to reorganizing a system to a different organizational structure. Most systems reorganize within the same structure, for example reorganizing in a hierarchy by changing the width or depth of the hierarchy. To approach this problem, we looked into adaptation of concepts and theories from social organization theory. In particular, we got insights from Schwaninger's model of Intelligent Human Organizations. We introduced a strategic reorganization model which enables the system to reorganize to a different type of organizational structure at run time. The proposed model employs different levels of organizational control for making organizational change decisions. We study the performance trade-offs and the efficacy of the proposed approach by running experiments using two instances of cooperative distributed problem solving applications. The results indicate that the proposed reorganization model results in performance improvements when task complexity increases

    Efficient Communication and Coordination for Large-Scale Multi-Agent Systems

    Get PDF
    The growth of the computational power of computers and the speed of networks has made large-scale multi-agent systems a promising technology. As the number of agents in a single application approaches thousands or millions, distributed computing has become a general paradigm in large-scale multi-agent systems to take the benefits of parallel computing. However, since these numerous agents are located on distributed computers and interact intensively with each other to achieve common goals, the agent communication cost significantly affects the performance of applications. Therefore, optimizing the agent communication cost on distributed systems could considerably reduce the runtime of multi-agent applications. Furthermore, because static multi-agent frameworks may not be suitable for all kinds of applications, and the communication patterns of agents may change during execution, multi-agent frameworks should adapt their services to support applications differently according to their dynamic characteristics. This thesis proposes three adaptive services at the agent framework level to reduce the agent communication and coordination cost of large-scale multi-agent applications. First, communication locality-aware agent distribution aims at minimizing inter-node communication by collocating heavily communicating agents on the same platform and maintaining agent group-based load sharing. Second, application agent-oriented middle agent services attempt to optimize agent interaction through middle agents by executing application agent-supported search algorithms on the middle agent address space. Third, message passing for mobile agents aims at reducing the time of message delivery to mobile agents using location caches or by extending the agent address scheme with location information. With these services, we have achieved very impressive experimental results in large- scale UAV simulations including up to 10,000 agents. Also, we have provided a formal definition of our framework and services with operational semantics
    corecore