1,071 research outputs found
Memory Structure and Cognitive Maps
A common way to understand memory structures in the cognitive sciences is as a cognitive map.
Cognitive maps are representational systems organized by dimensions shared with physical space. The
appeal to these maps begins literally: as an account of how spatial information is represented and used
to inform spatial navigation. Invocations of cognitive maps, however, are often more ambitious;
cognitive maps are meant to scale up and provide the basis for our more sophisticated memory
capacities. The extension is not meant to be metaphorical, but the way in which these richer mental
structures are supposed to remain map-like is rarely made explicit. Here we investigate this missing
link, asking: how do cognitive maps represent non-spatial information? We begin with a survey of
foundational work on spatial cognitive maps and then provide a comparative review of alternative,
non-spatial representational structures. We then turn to several cutting-edge projects that are engaged
in the task of scaling up cognitive maps so as to accommodate non-spatial information: first, on the
spatial-isometric approach , encoding content that is non-spatial but in some sense isomorphic to
spatial content; second, on the abstraction approach , encoding content that is an abstraction over
first-order spatial information; and third, on the embedding approach , embedding non-spatial
information within a spatial context, a prominent example being the Method-of-Loci. Putting these
cases alongside one another reveals the variety of options available for building cognitive maps, and the
distinctive limitations of each. We conclude by reflecting on where these results take us in terms of
understanding the place of cognitive maps in memory
Exploring the concept of interaction computing through the discrete algebraic analysis of the Belousov–Zhabotinsky reaction
Interaction computing (IC) aims to map the properties of integrable low-dimensional non-linear dynamical systems to the discrete domain of finite-state automata in an attempt to reproduce in software the self-organizing and dynamically stable properties of sub-cellular biochemical systems. As the work reported in this paper is still at the early stages of theory development it focuses on the analysis of a particularly simple chemical oscillator, the Belousov-Zhabotinsky (BZ) reaction. After retracing the rationale for IC developed over the past several years from the physical, biological, mathematical, and computer science points of view, the paper presents an elementary discussion of the Krohn-Rhodes decomposition of finite-state automata, including the holonomy decomposition of a simple automaton, and of its interpretation as an abstract positional number system. The method is then applied to the analysis of the algebraic properties of discrete finite-state automata derived from a simplified Petri net model of the BZ reaction. In the simplest possible and symmetrical case the corresponding automaton is, not surprisingly, found to contain exclusively cyclic groups. In a second, asymmetrical case, the decomposition is much more complex and includes five different simple non-abelian groups whose potential relevance arises from their ability to encode functionally complete algebras. The possible computational relevance of these findings is discussed and possible conclusions are drawn
Functional brain networks: great expectations, hard times and the big leap forward
Many physical and biological systems can be studied using complex network theory, a new statistical physics understanding of graph theory. The recent application of complex network theory to the study of functional brain networks has generated great enthusiasm as it allows addressing hitherto non-standard issues in the field, such as efficiency of brain functioning or vulnerability to damage. However, in spite of its high degree of generality, the theory was originally designed to describe systems profoundly different from the brain. We discuss some important caveats in the wholesale application of existing tools and concepts to a field they were not originally designed to describe. At the same time, we argue that complex network theory has not yet been taken full advantage of, as many of its important aspects are yet to make their appearance in the neuroscience literature. Finally, we propose that, rather than simply borrowing from an existing theory, functional neural networks can inspire a fundamental reformulation of complex network theory, to account for its exquisitely complex functioning mode
Digital Ecosystems: Ecosystem-Oriented Architectures
We view Digital Ecosystems to be the digital counterparts of biological
ecosystems. Here, we are concerned with the creation of these Digital
Ecosystems, exploiting the self-organising properties of biological ecosystems
to evolve high-level software applications. Therefore, we created the Digital
Ecosystem, a novel optimisation technique inspired by biological ecosystems,
where the optimisation works at two levels: a first optimisation, migration of
agents which are distributed in a decentralised peer-to-peer network, operating
continuously in time; this process feeds a second optimisation based on
evolutionary computing that operates locally on single peers and is aimed at
finding solutions to satisfy locally relevant constraints. The Digital
Ecosystem was then measured experimentally through simulations, with measures
originating from theoretical ecology, evaluating its likeness to biological
ecosystems. This included its responsiveness to requests for applications from
the user base, as a measure of the ecological succession (ecosystem maturity).
Overall, we have advanced the understanding of Digital Ecosystems, creating
Ecosystem-Oriented Architectures where the word ecosystem is more than just a
metaphor.Comment: 39 pages, 26 figures, journa
- …