46,275 research outputs found

    Decentralized dynamic task allocation for UAVs with limited communication range

    Full text link
    We present the Limited-range Online Routing Problem (LORP), which involves a team of Unmanned Aerial Vehicles (UAVs) with limited communication range that must autonomously coordinate to service task requests. We first show a general approach to cast this dynamic problem as a sequence of decentralized task allocation problems. Then we present two solutions both based on modeling the allocation task as a Markov Random Field to subsequently assess decisions by means of the decentralized Max-Sum algorithm. Our first solution assumes independence between requests, whereas our second solution also considers the UAVs' workloads. A thorough empirical evaluation shows that our workload-based solution consistently outperforms current state-of-the-art methods in a wide range of scenarios, lowering the average service time up to 16%. In the best-case scenario there is no gap between our decentralized solution and centralized techniques. In the worst-case scenario we manage to reduce by 25% the gap between current decentralized and centralized techniques. Thus, our solution becomes the method of choice for our problem

    Stabilizing Consensus with Many Opinions

    Full text link
    We consider the following distributed consensus problem: Each node in a complete communication network of size nn initially holds an \emph{opinion}, which is chosen arbitrarily from a finite set Σ\Sigma. The system must converge toward a consensus state in which all, or almost all nodes, hold the same opinion. Moreover, this opinion should be \emph{valid}, i.e., it should be one among those initially present in the system. This condition should be met even in the presence of an adaptive, malicious adversary who can modify the opinions of a bounded number of nodes in every round. We consider the \emph{3-majority dynamics}: At every round, every node pulls the opinion from three random neighbors and sets his new opinion to the majority one (ties are broken arbitrarily). Let kk be the number of valid opinions. We show that, if knαk \leqslant n^{\alpha}, where α\alpha is a suitable positive constant, the 3-majority dynamics converges in time polynomial in kk and logn\log n with high probability even in the presence of an adversary who can affect up to o(n)o(\sqrt{n}) nodes at each round. Previously, the convergence of the 3-majority protocol was known for Σ=2|\Sigma| = 2 only, with an argument that is robust to adversarial errors. On the other hand, no anonymous, uniform-gossip protocol that is robust to adversarial errors was known for Σ>2|\Sigma| > 2

    Deep Reinforcement Learning for Swarm Systems

    Full text link
    Recently, deep reinforcement learning (RL) methods have been applied successfully to multi-agent scenarios. Typically, these methods rely on a concatenation of agent states to represent the information content required for decentralized decision making. However, concatenation scales poorly to swarm systems with a large number of homogeneous agents as it does not exploit the fundamental properties inherent to these systems: (i) the agents in the swarm are interchangeable and (ii) the exact number of agents in the swarm is irrelevant. Therefore, we propose a new state representation for deep multi-agent RL based on mean embeddings of distributions. We treat the agents as samples of a distribution and use the empirical mean embedding as input for a decentralized policy. We define different feature spaces of the mean embedding using histograms, radial basis functions and a neural network learned end-to-end. We evaluate the representation on two well known problems from the swarm literature (rendezvous and pursuit evasion), in a globally and locally observable setup. For the local setup we furthermore introduce simple communication protocols. Of all approaches, the mean embedding representation using neural network features enables the richest information exchange between neighboring agents facilitating the development of more complex collective strategies.Comment: 31 pages, 12 figures, version 3 (published in JMLR Volume 20

    Simple Dynamics for Plurality Consensus

    Get PDF
    We study a \emph{Plurality-Consensus} process in which each of nn anonymous agents of a communication network initially supports an opinion (a color chosen from a finite set [k][k]). Then, in every (synchronous) round, each agent can revise his color according to the opinions currently held by a random sample of his neighbors. It is assumed that the initial color configuration exhibits a sufficiently large \emph{bias} ss towards a fixed plurality color, that is, the number of nodes supporting the plurality color exceeds the number of nodes supporting any other color by ss additional nodes. The goal is having the process to converge to the \emph{stable} configuration in which all nodes support the initial plurality. We consider a basic model in which the network is a clique and the update rule (called here the \emph{3-majority dynamics}) of the process is the following: each agent looks at the colors of three random neighbors and then applies the majority rule (breaking ties uniformly). We prove that the process converges in time O(min{k,(n/logn)1/3}logn)\mathcal{O}( \min\{ k, (n/\log n)^{1/3} \} \, \log n ) with high probability, provided that scmin{2k,(n/logn)1/3}nlogns \geqslant c \sqrt{ \min\{ 2k, (n/\log n)^{1/3} \}\, n \log n}. We then prove that our upper bound above is tight as long as k(n/logn)1/4k \leqslant (n/\log n)^{1/4}. This fact implies an exponential time-gap between the plurality-consensus process and the \emph{median} process studied by Doerr et al. in [ACM SPAA'11]. A natural question is whether looking at more (than three) random neighbors can significantly speed up the process. We provide a negative answer to this question: In particular, we show that samples of polylogarithmic size can speed up the process by a polylogarithmic factor only.Comment: Preprint of journal versio

    Bibliographic Review on Distributed Kalman Filtering

    Get PDF
    In recent years, a compelling need has arisen to understand the effects of distributed information structures on estimation and filtering. In this paper, a bibliographical review on distributed Kalman filtering (DKF) is provided.\ud The paper contains a classification of different approaches and methods involved to DKF. The applications of DKF are also discussed and explained separately. A comparison of different approaches is briefly carried out. Focuses on the contemporary research are also addressed with emphasis on the practical applications of the techniques. An exhaustive list of publications, linked directly or indirectly to DKF in the open literature, is compiled to provide an overall picture of different developing aspects of this area

    Dynamic max-consensus with local self-tuning

    Get PDF
    This work describes a novel control protocol for multi-agent systems to solve the dynamic max-consensus problem. In this problem, each agent has access to an external timevarying scalar signal and has the objective to estimate and track the maximum among all these signals by exploiting only local communications. The main strength of the proposed protocol is that it is able to self-tune its internal parameters in order to achieve an arbitrary small steady-state error without significantly affecting the convergence time. We employ the proposed protocol in the context of distributed graph parameter estimations, such as size, diameter, and radius, and provide simulations in the scenario of open multi-agent systems. Copyright (C) 2022 The Authors
    corecore