529 research outputs found
Dynamic Adaptive Computation: Tuning network states to task requirements
Neural circuits are able to perform computations under very diverse
conditions and requirements. The required computations impose clear constraints
on their fine-tuning: a rapid and maximally informative response to stimuli in
general requires decorrelated baseline neural activity. Such network dynamics
is known as asynchronous-irregular. In contrast, spatio-temporal integration of
information requires maintenance and transfer of stimulus information over
extended time periods. This can be realized at criticality, a phase transition
where correlations, sensitivity and integration time diverge. Being able to
flexibly switch, or even combine the above properties in a task-dependent
manner would present a clear functional advantage. We propose that cortex
operates in a "reverberating regime" because it is particularly favorable for
ready adaptation of computational properties to context and task. This
reverberating regime enables cortical networks to interpolate between the
asynchronous-irregular and the critical state by small changes in effective
synaptic strength or excitation-inhibition ratio. These changes directly adapt
computational properties, including sensitivity, amplification, integration
time and correlation length within the local network. We review recent
converging evidence that cortex in vivo operates in the reverberating regime,
and that various cortical areas have adapted their integration times to
processing requirements. In addition, we propose that neuromodulation enables a
fine-tuning of the network, so that local circuits can either decorrelate or
integrate, and quench or maintain their input depending on task. We argue that
this task-dependent tuning, which we call "dynamic adaptive computation",
presents a central organization principle of cortical networks and discuss
first experimental evidence.Comment: 6 pages + references, 2 figure
AI of Brain and Cognitive Sciences: From the Perspective of First Principles
Nowadays, we have witnessed the great success of AI in various applications,
including image classification, game playing, protein structure analysis,
language translation, and content generation. Despite these powerful
applications, there are still many tasks in our daily life that are rather
simple to humans but pose great challenges to AI. These include image and
language understanding, few-shot learning, abstract concepts, and low-energy
cost computing. Thus, learning from the brain is still a promising way that can
shed light on the development of next-generation AI. The brain is arguably the
only known intelligent machine in the universe, which is the product of
evolution for animals surviving in the natural environment. At the behavior
level, psychology and cognitive sciences have demonstrated that human and
animal brains can execute very intelligent high-level cognitive functions. At
the structure level, cognitive and computational neurosciences have unveiled
that the brain has extremely complicated but elegant network forms to support
its functions. Over years, people are gathering knowledge about the structure
and functions of the brain, and this process is accelerating recently along
with the initiation of giant brain projects worldwide. Here, we argue that the
general principles of brain functions are the most valuable things to inspire
the development of AI. These general principles are the standard rules of the
brain extracting, representing, manipulating, and retrieving information, and
here we call them the first principles of the brain. This paper collects six
such first principles. They are attractor network, criticality, random network,
sparse coding, relational memory, and perceptual learning. On each topic, we
review its biological background, fundamental property, potential application
to AI, and future development.Comment: 59 pages, 5 figures, review articl
Fractals in the Nervous System: conceptual Implications for Theoretical Neuroscience
This essay is presented with two principal objectives in mind: first, to
document the prevalence of fractals at all levels of the nervous system, giving
credence to the notion of their functional relevance; and second, to draw
attention to the as yet still unresolved issues of the detailed relationships
among power law scaling, self-similarity, and self-organized criticality. As
regards criticality, I will document that it has become a pivotal reference
point in Neurodynamics. Furthermore, I will emphasize the not yet fully
appreciated significance of allometric control processes. For dynamic fractals,
I will assemble reasons for attributing to them the capacity to adapt task
execution to contextual changes across a range of scales. The final Section
consists of general reflections on the implications of the reviewed data, and
identifies what appear to be issues of fundamental importance for future
research in the rapidly evolving topic of this review
The emergence of synaesthesia in a Neuronal Network Model via changes in perceptual sensitivity and plasticity
Synaesthesia is an unusual perceptual experience in which an inducer stimulus triggers a percept in a different domain in addition to its own. To explore the conditions under which synaesthesia evolves, we studied a neuronal network model that represents two recurrently connected neural systems. The interactions in the network evolve according to learning rules that optimize sensory sensitivity. We demonstrate several scenarios, such as sensory deprivation or heightened plasticity, under which synaesthesia can evolve even though the inputs to the two systems are statistically independent and the initial cross-talk interactions are zero. Sensory deprivation is the known causal mechanism for acquired synaesthesia and increased plasticity is implicated in developmental synaesthesia. The model unifies different causes of synaesthesia within a single theoretical framework and repositions synaesthesia not as some quirk of aberrant connectivity, but rather as a functional brain state that can emerge as a consequence of optimising sensory information processing
Stimulus sensitivity of a spiking neural network model
Some recent papers relate the criticality of complex systems to their maximal
capacity of information processing. In the present paper, we consider high
dimensional point processes, known as age-dependent Hawkes processes, which
have been used to model spiking neural networks. Using mean-field
approximation, the response of the network to a stimulus is computed and we
provide a notion of stimulus sensitivity. It appears that the maximal
sensitivity is achieved in the sub-critical regime, yet almost critical for a
range of biologically relevant parameters
Can biological quantum networks solve NP-hard problems?
There is a widespread view that the human brain is so complex that it cannot
be efficiently simulated by universal Turing machines. During the last decades
the question has therefore been raised whether we need to consider quantum
effects to explain the imagined cognitive power of a conscious mind.
This paper presents a personal view of several fields of philosophy and
computational neurobiology in an attempt to suggest a realistic picture of how
the brain might work as a basis for perception, consciousness and cognition.
The purpose is to be able to identify and evaluate instances where quantum
effects might play a significant role in cognitive processes.
Not surprisingly, the conclusion is that quantum-enhanced cognition and
intelligence are very unlikely to be found in biological brains. Quantum
effects may certainly influence the functionality of various components and
signalling pathways at the molecular level in the brain network, like ion
ports, synapses, sensors, and enzymes. This might evidently influence the
functionality of some nodes and perhaps even the overall intelligence of the
brain network, but hardly give it any dramatically enhanced functionality. So,
the conclusion is that biological quantum networks can only approximately solve
small instances of NP-hard problems.
On the other hand, artificial intelligence and machine learning implemented
in complex dynamical systems based on genuine quantum networks can certainly be
expected to show enhanced performance and quantum advantage compared with
classical networks. Nevertheless, even quantum networks can only be expected to
efficiently solve NP-hard problems approximately. In the end it is a question
of precision - Nature is approximate.Comment: 38 page
Optimal Input Representation in Neural Systems at the Edge of Chaos
Shedding light on how biological systems represent, process and store information in noisy
environments is a key and challenging goal. A stimulating, though controversial, hypothesis poses
that operating in dynamical regimes near the edge of a phase transition, i.e., at criticality or the âedge
of chaosâ, can provide information-processing living systems with important operational advantages,
creating, e.g., an optimal trade-off between robustness and flexibility. Here, we elaborate on a recent
theoretical result, which establishes that the spectrum of covariance matrices of neural networks
representing complex inputs in a robust way needs to decay as a power-law of the rank, with an
exponent close to unity, a result that has been indeed experimentally verified in neurons of the mouse
visual cortex. Aimed at understanding and mimicking these results, we construct an artificial neural
network and train it to classify images. We find that the best performance in such a task is obtained
when the network operates near the critical point, at which the eigenspectrum of the covariance
matrix follows the very same statistics as actual neurons do. Thus, we conclude that operating near
criticality can also haveâbesides the usually alleged virtuesâthe advantage of allowing for flexible,
robust and efficient input representations.The Spanish Ministry and Agencia Estatal de investigaciĂłn
(AEI) through grant FIS2017-84256-P (European Regional Development Fund)âConsejerĂa de Conocimiento, InvestigaciĂłn Universidad, Junta de AndalucĂaâ and European Regional
Development Fund, Project Ref. A-FQM-175-UGR18 and Project Ref. P20-0017
Coordination dynamics in the sensorimotor loop
The last two decades have witnessed radical changes of perspective about the nature of intelligence and cognition, leaving behind some of the assumptions of computational functionalism. From the myriad of approaches seeking to substitute the old rule-based symbolic perception of mind, we are especially interested in two of them. The first is Embodied and Situated Cognition, where the advances in modeling complex adaptive systems through computer simulations have reconfigured the way in which mechanistic, embodied and interactive explanations can conceptualize the mind. We are particularly interested in the concept of sensorimotor loop, which brings a new perspective about what is needed for a meaningful interaction with the environment, emphasizing the role of the coordination of effector and sensor activities while performing a concrete task. The second one is the framework of Coordination Dynamics, which has been developed as a result of the increasing focus of neuroscience on self-organized oscillatory brain dynamics. It provides formal tools to study the mechanisms through which complex biological systems stabilize coordination states under conditions in which they would otherwise become unstable. We will merge both approaches and define coordination in the sensorimotor loop as the main phenomena behind the emergence of cognitive behavior. At the same time, we will provide methodological tools and concepts to address this hypothesis. Finally, we will present two case studies based on the proposed approach: 1. We will study the phenomenon known as âintermittent behaviorâ, which is observed in organisms at different levels (from microorganisms to higher animals). We will propose a model that understands intermittent behavior as a general strategy of biologica organization when an organism has to adapt to complex changing environments, and would allow to establish effective sensorimotor loops even in situations of instable engagement with the world. 2. We will perform a simulation of a phonotaxis task performed by an agent with an oscillator network as neural controller. The objective will be to characterize robust adaptive coupling between perceptive activity and the environmental dynamics just through phase information processing. We will observe how the robustness of the coupling crucially depends of how the sensorimotor loop structures and constrains both the emergent neural and behavioral patterns. We will hypothesize that this structuration of the sensorimotor space, in which only meaningful behavioral patterns can be stabilized, is a key ingredient for the emergence of higher cognitive abilities
- âŠ