363 research outputs found

    Neural Avalanches at the Critical Point between Replay and Non-Replay of Spatiotemporal Patterns

    Get PDF
    We model spontaneous cortical activity with a network of coupled spiking units, in which multiple spatio-temporal patterns are stored as dynamical attractors. We introduce an order parameter, which measures the overlap (similarity) between the activity of the network and the stored patterns. We find that, depending on the excitability of the network, different working regimes are possible. For high excitability, the dynamical attractors are stable, and a collective activity that replays one of the stored patterns emerges spontaneously, while for low excitability, no replay is induced. Between these two regimes, there is a critical region in which the dynamical attractors are unstable, and intermittent short replays are induced by noise. At the critical spiking threshold, the order parameter goes from zero to one, and its fluctuations are maximized, as expected for a phase transition (and as observed in recent experimental results in the brain). Notably, in this critical region, the avalanche size and duration distributions follow power laws. Critical exponents are consistent with a scaling relationship observed recently in neural avalanches measurements. In conclusion, our simple model suggests that avalanche power laws in cortical spontaneous activity may be the effect of a network at the critical point between the replay and non-replay of spatio-temporal patterns

    Corticonic models of brain mechanisms underlying cognition and intelligence

    Get PDF
    The concern of this review is brain theory or more specifically, in its first part, a model of the cerebral cortex and the way it:(a) interacts with subcortical regions like the thalamus and the hippocampus to provide higher-level-brain functions that underlie cognition and intelligence, (b) handles and represents dynamical sensory patterns imposed by a constantly changing environment, (c) copes with the enormous number of such patterns encountered in a lifetime bymeans of dynamic memory that offers an immense number of stimulus-specific attractors for input patterns (stimuli) to select from, (d) selects an attractor through a process of “conjugation” of the input pattern with the dynamics of the thalamo–cortical loop, (e) distinguishes between redundant (structured)and non-redundant (random) inputs that are void of information, (f) can do categorical perception when there is access to vast associative memory laid out in the association cortex with the help of the hippocampus, and (g) makes use of “computation” at the edge of chaos and information driven annealing to achieve all this. Other features and implications of the concepts presented for the design of computational algorithms and machines with brain-like intelligence are also discussed. The material and results presented suggest, that a Parametrically Coupled Logistic Map network (PCLMN) is a minimal model of the thalamo–cortical complex and that marrying such a network to a suitable associative memory with re-entry or feedback forms a useful, albeit, abstract model of a cortical module of the brain that could facilitate building a simple artificial brain. In the second part of the review, the results of numerical simulations and drawn conclusions in the first part are linked to the most directly relevant works and views of other workers. What emerges is a picture of brain dynamics on the mesoscopic and macroscopic scales that gives a glimpse of the nature of the long sought after brain code underlying intelligence and other higher level brain functions. Physics of Life Reviews 4 (2007) 223–252 © 2007 Elsevier B.V. All rights reserved

    Fractals in the Nervous System: conceptual Implications for Theoretical Neuroscience

    Get PDF
    This essay is presented with two principal objectives in mind: first, to document the prevalence of fractals at all levels of the nervous system, giving credence to the notion of their functional relevance; and second, to draw attention to the as yet still unresolved issues of the detailed relationships among power law scaling, self-similarity, and self-organized criticality. As regards criticality, I will document that it has become a pivotal reference point in Neurodynamics. Furthermore, I will emphasize the not yet fully appreciated significance of allometric control processes. For dynamic fractals, I will assemble reasons for attributing to them the capacity to adapt task execution to contextual changes across a range of scales. The final Section consists of general reflections on the implications of the reviewed data, and identifies what appear to be issues of fundamental importance for future research in the rapidly evolving topic of this review

    Trajectory prediction of moving objects by means of neural networks

    Get PDF
    Thesis (Master)--Izmir Institute of Technology, Computer Engineering, Izmir, 1997Includes bibliographical references (leaves: 103-105)Text in English; Abstract: Turkish and Englishviii, 105 leavesEstimating the three-dimensional motion of an object from a sequence of object positions and orientation is of significant importance in variety of applications in control and robotics. For instance, autonomous navigation, manipulation, servo, tracking, planning and surveillance needs prediction of motion parameters. Although "motion estimation" is an old problem (the formulations date back to the beginning of the century), only recently scientists have provided with the tools from nonlinear system estimation theory to solve this problem eural Networks are the ones which have recently been used in many nonlinear dynamic system parameter estimation context. The approximating ability of the neural network is used to identifY the relation between system variables and parameters of a dynamic system. The position, velocity and acceleration of the object are estimated by several neural networks using the II most recent measurements of the object coordinates as input to the system Several neural network topologies with different configurations are introduced and utilized in the solution of the problem. Training schemes for each configuration are given in detail. Simulation results for prediction of motion having different characteristics via different architectures with alternative configurations are presented comparatively

    Linear and nonlinear approaches to unravel dynamics and connectivity in neuronal cultures

    Get PDF
    [eng] In the present thesis, we propose to explore neuronal circuits at the mesoscale, an approach in which one monitors small populations of few thousand neurons and concentrates in the emergence of collective behavior. In our case, we carried out such an exploration both experimentally and numerically, and by adopting an analysis perspective centered on time series analysis and dynamical systems. Experimentally, we used neuronal cultures and prepared more than 200 of them, which were monitored using fluorescence calcium imaging. By adjusting the experimental conditions, we could set two basic arrangements of neurons, namely homogeneous and aggregated. In the experiments, we carried out two major explorations, namely development and disintegration. In the former we investigated changes in network behavior as it matured; in the latter we applied a drug that reduced neuronal interconnectivity. All the subsequent analyses and modeling along the thesis are based on these experimental data. Numerically, the thesis comprised two aspects. The first one was oriented towards a simulation of neuronal connectivity and dynamics. The second one was oriented towards the development of linear and nonlinear analysis tools to unravel dynamic and connectivity aspects of the measured experimental networks. For the first aspect, we developed a sophisticated software package to simulate single neuronal dynamics using a quadratic integrate–and–fire model with adaptation and depression. This model was plug into a synthetic graph in which the nodes of the network are neurons, and the edges connections. The graph was created using spatial embedding and realistic biology. We carried out hundreds of simulations in which we tuned the density of neurons, their spatial arrangement and the characteristics of the fluorescence signal. As a key result, we observed that homogeneous networks required a substantial number of neurons to fire and exhibit collective dynamics, and that the presence of aggregation significantly reduced the number of required neurons. For the second aspect, data analysis, we analyzed experiments and simulations to tackle three major aspects: network dynamics reconstruction using linear descriptions, dynamics reconstruction using nonlinear descriptors, and the assessment of neuronal connectivity from solely activity data. For the linear study, we analyzed all experiments using the power spectrum density (PSD), and observed that it was sufficiently good to describe the development of the network or its disintegration. PSD also allowed us to distinguish between healthy and unhealthy networks, and revealed dynamical heterogeneities across the network. For the nonlinear study, we used techniques in the context of recurrence plots. We first characterized the embedding dimension m and the time delay δ for each experiment, built the respective recurrence plots, and extracted key information of the dynamics of the system through different descriptors. Experimental results were contrasted with numerical simulations. After analyzing about 400 time series, we concluded that the degree of dynamical complexity in neuronal cultures changes both during development and disintegration. We also observed that the healthier the culture, the higher its dynamic complexity. Finally, for the reconstruction study, we first used numerical simulations to determine the best measure of ‘statistical interdependence’ among any two neurons, and took Generalized Transfer Entropy. We then analyzed the experimental data. We concluded that young cultures have a weak connectivity that increases along maturation. Aggregation increases average connectivity, and more interesting, also the assortativity, i.e. the tendency of highly connected nodes to connect with other highly connected node. In turn, this assortativity may delineates important aspects of the dynamics of the network. Overall, the results show that spatial arrangement and neuronal dynamics are able to shape a very rich repertoire of dynamical states of varying complexity.[cat] L’habilitat dels teixits neuronals de processar i transmetre informació de forma eficient depèn de les propietats dinàmiques intrínseques de les neurones i de la connectivitat entre elles. La present tesi proposa explorar diferents tècniques experimentals i de simulació per analitzar la dinàmica i connectivitat de xarxes neuronals corticals de rata embrionària. Experimentalment, la gravació de l’activitat espontània d’una població de neurones en cultiu, mitjançant una càmera ràpida i tècniques de fluorescència, possibilita el seguiment de forma controlada de l’activitat individual de cada neurona, així com la modificació de la seva connectivitat. En conjunt, aquestes eines permeten estudiar el comportament col.lectiu emergent de la població neuronal. Amb l’objectiu de simular els patrons observats en el laboratori, hem implementat un model mètric aleatori de creixement neuronal per simular la xarxa física de connexions entre neurones, i un model quadràtic d’integració i dispar amb adaptació i depressió per modelar l’ampli espectre de dinàmiques neuronals amb un cost computacional reduït. Hem caracteritzat la dinàmica global i individual de les neurones i l’hem correlacionat amb la seva estructura subjacent mitjançant tècniques lineals i no–lineals de series temporals. L’anàlisi espectral ens ha possibilitat la descripció del desenvolupament i els canvis en connectivitat en els cultius, així com la diferenciació entre cultius sans dels patològics. La reconstrucció de la dinàmica subjacent mitjançant mètodes d’incrustació i l’ús de gràfics de recurrència ens ha permès detectar diferents transicions dinàmiques amb el corresponent guany o pèrdua de la complexitat i riquesa dinàmica del cultiu durant els diferents estudis experimentals. Finalment, a fi de reconstruir la connectivitat interna hem testejat, mitjançant simulacions, diferents quantificadors per mesurar la dependència estadística entre neurona i neurona, seleccionant finalment el mètode de transferència d’entropia gereralitzada. Seguidament, hem procedit a caracteritzar les xarxes amb diferents paràmetres. Malgrat presentar certs tres de xarxes tipus ‘petit món’, els nostres cultius mostren una distribució de grau ‘exponencial’ o ‘esbiaixada’ per, respectivament, cultius joves i madurs. Addicionalment, hem observat que les xarxes homogènies presenten la propietat de disassortativitat, mentre que xarxes amb un creixent nivell d’agregació espaial presenten assortativitat. Aquesta propietat impacta fortament en la transmissió, resistència i sincronització de la xarxa

    Random Recurrent Neural Networks Dynamics

    Full text link
    This paper is a review dealing with the study of large size random recurrent neural networks. The connection weights are selected according to a probability law and it is possible to predict the network dynamics at a macroscopic scale using an averaging principle. After a first introductory section, the section 1 reviews the various models from the points of view of the single neuron dynamics and of the global network dynamics. A summary of notations is presented, which is quite helpful for the sequel. In section 2, mean-field dynamics is developed. The probability distribution characterizing global dynamics is computed. In section 3, some applications of mean-field theory to the prediction of chaotic regime for Analog Formal Random Recurrent Neural Networks (AFRRNN) are displayed. The case of AFRRNN with an homogeneous population of neurons is studied in section 4. Then, a two-population model is studied in section 5. The occurrence of a cyclo-stationary chaos is displayed using the results of \cite{Dauce01}. In section 6, an insight of the application of mean-field theory to IF networks is given using the results of \cite{BrunelHakim99}.Comment: Review paper, 36 pages, 5 figure
    corecore