94,703 research outputs found
Distributed Constraint Optimization Problems and Applications: A Survey
The field of Multi-Agent System (MAS) is an active area of research within
Artificial Intelligence, with an increasingly important impact in industrial
and other real-world applications. Within a MAS, autonomous agents interact to
pursue personal interests and/or to achieve common objectives. Distributed
Constraint Optimization Problems (DCOPs) have emerged as one of the prominent
agent architectures to govern the agents' autonomous behavior, where both
algorithms and communication models are driven by the structure of the specific
problem. During the last decade, several extensions to the DCOP model have
enabled them to support MAS in complex, real-time, and uncertain environments.
This survey aims at providing an overview of the DCOP model, giving a
classification of its multiple extensions and addressing both resolution
methods and applications that find a natural mapping within each class of
DCOPs. The proposed classification suggests several future perspectives for
DCOP extensions, and identifies challenges in the design of efficient
resolution algorithms, possibly through the adaptation of strategies from
different areas
Multi-Agent Distributed Lifelong Learning for Collective Knowledge Acquisition
Lifelong machine learning methods acquire knowledge over a series of
consecutive tasks, continually building upon their experience. Current lifelong
learning algorithms rely upon a single learning agent that has centralized
access to all data. In this paper, we extend the idea of lifelong learning from
a single agent to a network of multiple agents that collectively learn a series
of tasks. Each agent faces some (potentially unique) set of tasks; the key idea
is that knowledge learned from these tasks may benefit other agents trying to
learn different (but related) tasks. Our Collective Lifelong Learning Algorithm
(CoLLA) provides an efficient way for a network of agents to share their
learned knowledge in a distributed and decentralized manner, while preserving
the privacy of the locally observed data. Note that a decentralized scheme is a
subclass of distributed algorithms where a central server does not exist and in
addition to data, computations are also distributed among the agents. We
provide theoretical guarantees for robust performance of the algorithm and
empirically demonstrate that CoLLA outperforms existing approaches for
distributed multi-task learning on a variety of data sets
Multi Objective Particle Swarm Optimization based Cooperative Agents with Automated Negotiation
This paper investigates a new hybridization of multi-objective particle swarm
optimization (MOPSO) and cooperative agents (MOPSO-CA) to handle the problem of
stagnation encounters in MOPSO, which leads solutions to trap in local optima.
The proposed approach involves a new distribution strategy based on the idea of
having a set of a sub-population, each of which is processed by one agent. The
number of the sub-population and agents are adjusted dynamically through the
Pareto ranking. This method allocates a dynamic number of sub-population as
required to improve diversity in the search space. Additionally, agents are
used for better management for the exploitation within a sub-population, and
for exploration among sub-populations. Furthermore, we investigate the
automated negotiation within agents in order to share the best knowledge. To
validate our approach, several benchmarks are performed. The results show that
the introduced variant ensures the trade-off between the exploitation and
exploration with respect to the comparative algorithm
Multi-Cycle Assignment Problems with Rotational Diversity
Multi-cycle assignment problems address scenarios where a series of general
assignment problems has to be solved sequentially. Subsequent cycles can differ
from previous ones due to changing availability or creation of tasks and
agents, which makes an upfront static schedule infeasible and introduces
uncertainty in the task-agent assignment process. We consider the setting
where, besides profit maximization, it is also desired to maintain diverse
assignments for tasks and agents, such that all tasks have been assigned to all
agents over subsequent cycles. This problem of multi-cycle assignment with
rotational diversity is approached in two sub-problems: The outer problem which
augments the original profit maximization objective with additional information
about the state of rotational diversity while the inner problem solves the
adjusted general assignment problem in a single execution of the model. We
discuss strategies to augment the profit values and evaluate them
experimentally. The method's efficacy is shown in three case studies:
multi-cycle variants of the multiple knapsack and the multiple subset sum
problems, and a real-world case study on the test case selection and assignment
problem from the software engineering domain.Comment: Extended journal versio
Autonomous Wireless Systems with Artificial Intelligence
This paper discusses technology and opportunities to embrace artificial
intelligence (AI) in the design of autonomous wireless systems. We aim to
provide readers with motivation and general AI methodology of autonomous agents
in the context of self-organization in real time by unifying knowledge
management with sensing, reasoning and active learning. We highlight
differences between training-based methods for matching problems and
training-free methods for environment-specific problems. Finally, we
conceptually introduce the functions of an autonomous agent with knowledge
management
Distributed MIN-MAX Optimization Application to Time-optimal Consensus: An Alternating Projection Approach
In this paper, we proposed an alternating projection based algorithm to solve
a class of distributed MIN-MAX convex optimization problems. We firstly
transform this MINMAX problem into the problem of searching for the minimum
distance between some hyper-plane and the intersection of the epigraphs of
convex functions. The Bregman's alternating method is employed in our algorithm
to achieve the distance by iteratively projecting onto the hyper-plane and the
intersection. The projection onto the intersection is obtained by cyclic
Dykstra's projection method. We further apply our algorithm to the minimum time
multi-agent consensus problem. The attainable states set for the agent can be
transformed into the epigraph of some convex functions, and the search for
time-optimal state for consensus satisfies the MIN-MAX problem formulation.
Finally, the numerous simulation proves the validity of our algorithm.Comment: 11 pages, 6 figures, submitted to AIAA GNC 201
Partially Observed, Multi-objective Markov Games
The intent of this research is to generate a set of non-dominated policies
from which one of two agents (the leader) can select a most preferred policy to
control a dynamic system that is also affected by the control decisions of the
other agent (the follower). The problem is described by an infinite horizon,
partially observed Markov game (POMG). At each decision epoch, each agent
knows: its past and present states, its past actions, and noise corrupted
observations of the other agent's past and present states. The actions of each
agent are determined at each decision epoch based on these data. The leader
considers multiple objectives in selecting its policy. The follower considers a
single objective in selecting its policy with complete knowledge of and in
response to the policy selected by the leader. This leader-follower assumption
allows the POMG to be transformed into a specially structured, partially
observed Markov decision process (POMDP). This POMDP is used to determine the
follower's best response policy. A multi-objective genetic algorithm (MOGA) is
used to create the next generation of leader policies based on the fitness
measures of each leader policy in the current generation. Computing a fitness
measure for a leader policy requires a value determination calculation, given
the leader policy and the follower's best response policy. The policies from
which the leader can select a most preferred policy are the non-dominated
policies of the final generation of leader policies created by the MOGA. An
example is presented that illustrates how these results can be used to support
a manager of a liquid egg production process (the leader) in selecting a
sequence of actions to best control this process over time, given that there is
an attacker (the follower) who seeks to contaminate the liquid egg production
process with a chemical or biological toxin
A Survey and Critique of Multiagent Deep Reinforcement Learning
Deep reinforcement learning (RL) has achieved outstanding results in recent
years. This has led to a dramatic increase in the number of applications and
methods. Recent works have explored learning beyond single-agent scenarios and
have considered multiagent learning (MAL) scenarios. Initial results report
successes in complex multiagent domains, although there are several challenges
to be addressed. The primary goal of this article is to provide a clear
overview of current multiagent deep reinforcement learning (MDRL) literature.
Additionally, we complement the overview with a broader analysis: (i) we
revisit previous key components, originally presented in MAL and RL, and
highlight how they have been adapted to multiagent deep reinforcement learning
settings. (ii) We provide general guidelines to new practitioners in the area:
describing lessons learned from MDRL works, pointing to recent benchmarks, and
outlining open avenues of research. (iii) We take a more critical tone raising
practical challenges of MDRL (e.g., implementation and computational demands).
We expect this article will help unify and motivate future research to take
advantage of the abundant literature that exists (e.g., RL and MAL) in a joint
effort to promote fruitful research in the multiagent community.Comment: Under review since Oct 2018. Earlier versions of this work had the
title: "Is multiagent deep reinforcement learning the answer or the question?
A brief survey
A Unifying Survey of Reinforced, Sensitive and Stigmergic Agent-Based Approaches for E-GTSP
The Generalized Traveling Salesman Problem (GTSP) is one of the NP-hard
combinatorial optimization problems. A variant of GTSP is E-GTSP where E,
meaning equality, has the constraint: exactly one node from a cluster of a
graph partition is visited. The main objective of the E-GTSP is to find a
minimum cost tour passing through exactly one node from each cluster of an
undirected graph. Agent-based approaches involving are successfully used
nowadays for solving real life complex problems. The aim of the current paper
is to illustrate some variants of agent-based algorithms including ant-based
models with specific properties for solving E-GTSP.Comment: 9 pages, 2 figure
A Survey on Traffic Signal Control Methods
Traffic signal control is an important and challenging real-world problem,
which aims to minimize the travel time of vehicles by coordinating their
movements at the road intersections. Current traffic signal control systems in
use still rely heavily on oversimplified information and rule-based methods,
although we now have richer data, more computing power and advanced methods to
drive the development of intelligent transportation. With the growing interest
in intelligent transportation using machine learning methods like reinforcement
learning, this survey covers the widely acknowledged transportation approaches
and a comprehensive list of recent literature on reinforcement for traffic
signal control. We hope this survey can foster interdisciplinary research on
this important topic.Comment: 32 page
- …