5,220 research outputs found
Multilayer Networks
In most natural and engineered systems, a set of entities interact with each
other in complicated patterns that can encompass multiple types of
relationships, change in time, and include other types of complications. Such
systems include multiple subsystems and layers of connectivity, and it is
important to take such "multilayer" features into account to try to improve our
understanding of complex systems. Consequently, it is necessary to generalize
"traditional" network theory by developing (and validating) a framework and
associated tools to study multilayer systems in a comprehensive fashion. The
origins of such efforts date back several decades and arose in multiple
disciplines, and now the study of multilayer networks has become one of the
most important directions in network science. In this paper, we discuss the
history of multilayer networks (and related concepts) and review the exploding
body of work on such networks. To unify the disparate terminology in the large
body of recent work, we discuss a general framework for multilayer networks,
construct a dictionary of terminology to relate the numerous existing concepts
to each other, and provide a thorough discussion that compares, contrasts, and
translates between related notions such as multilayer networks, multiplex
networks, interdependent networks, networks of networks, and many others. We
also survey and discuss existing data sets that can be represented as
multilayer networks. We review attempts to generalize single-layer-network
diagnostics to multilayer networks. We also discuss the rapidly expanding
research on multilayer-network models and notions like community structure,
connected components, tensor decompositions, and various types of dynamical
processes on multilayer networks. We conclude with a summary and an outlook.Comment: Working paper; 59 pages, 8 figure
A Tale of Two Animats: What does it take to have goals?
What does it take for a system, biological or not, to have goals? Here, this
question is approached in the context of in silico artificial evolution. By
examining the informational and causal properties of artificial organisms
('animats') controlled by small, adaptive neural networks (Markov Brains), this
essay discusses necessary requirements for intrinsic information, autonomy, and
meaning. The focus lies on comparing two types of Markov Brains that evolved in
the same simple environment: one with purely feedforward connections between
its elements, the other with an integrated set of elements that causally
constrain each other. While both types of brains 'process' information about
their environment and are equally fit, only the integrated one forms a causally
autonomous entity above a background of external influences. This suggests that
to assess whether goals are meaningful for a system itself, it is important to
understand what the system is, rather than what it does.Comment: This article is a contribution to the FQXi 2016-2017 essay contest
"Wandering Towards a Goal
Achievable bisimilar behaviour of abstract state systems
Given a plant system and a desired system, we study conditions for which there exists a controller that interconnected with the plant, yields a system that is bisimilar to the desired system. Some sufficient and some necessary conditions are provided in the general case of (non-deterministic) abstract state systems and stronger results are obtained for the special classes of autonomous abstract state systems, finite abstract state systems, and non-deterministic linear dynamical systems
Statistical Physics and Representations in Real and Artificial Neural Networks
This document presents the material of two lectures on statistical physics
and neural representations, delivered by one of us (R.M.) at the Fundamental
Problems in Statistical Physics XIV summer school in July 2017. In a first
part, we consider the neural representations of space (maps) in the
hippocampus. We introduce an extension of the Hopfield model, able to store
multiple spatial maps as continuous, finite-dimensional attractors. The phase
diagram and dynamical properties of the model are analyzed. We then show how
spatial representations can be dynamically decoded using an effective Ising
model capturing the correlation structure in the neural data, and compare
applications to data obtained from hippocampal multi-electrode recordings and
by (sub)sampling our attractor model. In a second part, we focus on the problem
of learning data representations in machine learning, in particular with
artificial neural networks. We start by introducing data representations
through some illustrations. We then analyze two important algorithms, Principal
Component Analysis and Restricted Boltzmann Machines, with tools from
statistical physics
- …