20 research outputs found
The Benefits of Interaction Constraints in Distributed Autonomous Systems
The design of distributed autonomous systems often omits consideration of the
underlying network dynamics. Recent works in multi-agent systems and swarm
robotics alike have highlighted the impact that the interactions between agents
have on the collective behaviours exhibited by the system. In this paper, we
seek to highlight the role that the underlying interaction network plays in
determining the performance of the collective behaviour of a system, comparing
its impact with that of the physical network. We contextualise this by defining
a collective learning problem in which agents must reach a consensus about
their environment in the presence of noisy information. We show that the
physical connectivity of the agents plays a less important role than when an
interaction network of limited connectivity is imposed on the system to
constrain agent communication. Constraining agent interactions in this way
drastically improves the performance of the system in a collective learning
context. Additionally, we provide further evidence for the idea that `less is
more' when it comes to propagating information in distributed autonomous
systems for the purpose of collective learning.Comment: To appear in the Proceedings of the Distributed Autonomous Robotic
Systems 16th International Symposium (2022
The Impact of Network Connectivity on Collective Learning
In decentralised autonomous systems it is the interactions between individual
agents which govern the collective behaviours of the system. These local-level
interactions are themselves often governed by an underlying network structure.
These networks are particularly important for collective learning and
decision-making whereby agents must gather evidence from their environment and
propagate this information to other agents in the system. Models for collective
behaviours may often rely upon the assumption of total connectivity between
agents to provide effective information sharing within the system, but this
assumption may be ill-advised. In this paper we investigate the impact that the
underlying network has on performance in the context of collective learning.
Through simulations we study small-world networks with varying levels of
connectivity and randomness and conclude that totally-connected networks result
in higher average error when compared to networks with less connectivity.
Furthermore, we show that networks of high regularity outperform networks with
increasing levels of random connectivity.Comment: 13 pages, 5 figures. To appear at the 15th International Symposium on
Distributed Autonomous Robotic Systems 2021. Presented at the joint
DARS-SWARM 2021 symposium held (virtually) in Kyoto, Japa
A model of multi-agent consensus for vague and uncertain beliefs
Consensus formation is investigated for multi-agent systems in which agents’ beliefs are both vague and uncertain. Vagueness is represented by a third truth state meaning borderline. This is combined with a probabilistic model of uncertainty. A belief combination operator is then proposed, which exploits borderline truth values to enable agents with conflicting beliefs to reach a compromise. A number of simulation experiments are carried out, in which agents apply this operator in pairwise interactions, under the bounded confidence restriction that the two agents’ beliefs must be sufficiently consistent with each other before agreement can be reached. As well as studying the consensus operator in isolation, we also investigate scenarios in which agents are influenced either directly or indirectly by the state of the world. For the former, we conduct simulations that combine consensus formation with belief updating based on evidence. For the latter, we investigate the effect of assuming that the closer an agent’s beliefs are to the truth the more visible they are in the consensus building process. In all cases, applying the consensus operators results in the population converging to a single shared belief that is both crisp and certain. Furthermore, simulations that combine consensus formation with evidential updating converge more quickly to a shared opinion, which is closer to the actual state of the world than those in which beliefs are only changed as a result of directly receiving new evidence. Finally, if agent interactions are guided by belief quality measured as similarity to the true state of the world, then applying the consensus operator alone results in the population converging to a high-quality shared belief
On the Existence of Information Bottlenecks in Living and Non-Living Systems
In many complex systems, we observe that `interesting behaviour' is often the
consequence of a system exploiting the existence of an Information Bottleneck
(IB). These bottlenecks can occur at different scales, between individuals or
components of a system, and sometimes within individuals themselves.
Oftentimes, we regard these bottlenecks negatively; as merely the limitations
of the individual's physiology and something that ought to be overcome when
designing and implementing artificial systems. However, we suggest instead that
IBs may serve a purpose beyond merely providing a minimally-viable channel for
coordination in collective systems. More specifically, we suggest that
interesting or novel behaviour occurs when the individuals in a system are
constrained or limited in their ability to share information and must discover
novel ways to exploit existing mechanisms, which are inherently bottlenecked,
rather than circumventing or otherwise avoiding those mechanisms entirely.Comment: To appear in the proceedings of the 2023 Conference on Artificial
Lif