30 research outputs found

    Spherical and Hyperbolic Toric Topology-Based Codes On Graph Embedding for Ising MRF Models: Classical and Quantum Topology Machine Learning

    Full text link
    The paper introduces the application of information geometry to describe the ground states of Ising models by utilizing parity-check matrices of cyclic and quasi-cyclic codes on toric and spherical topologies. The approach establishes a connection between machine learning and error-correcting coding. This proposed approach has implications for the development of new embedding methods based on trapping sets. Statistical physics and number geometry applied for optimize error-correcting codes, leading to these embedding and sparse factorization methods. The paper establishes a direct connection between DNN architecture and error-correcting coding by demonstrating how state-of-the-art architectures (ChordMixer, Mega, Mega-chunk, CDIL, ...) from the long-range arena can be equivalent to of block and convolutional LDPC codes (Cage-graph, Repeat Accumulate). QC codes correspond to certain types of chemical elements, with the carbon element being represented by the mixed automorphism Shu-Lin-Fossorier QC-LDPC code. The connections between Belief Propagation and the Permanent, Bethe-Permanent, Nishimori Temperature, and Bethe-Hessian Matrix are elaborated upon in detail. The Quantum Approximate Optimization Algorithm (QAOA) used in the Sherrington-Kirkpatrick Ising model can be seen as analogous to the back-propagation loss function landscape in training DNNs. This similarity creates a comparable problem with TS pseudo-codeword, resembling the belief propagation method. Additionally, the layer depth in QAOA correlates to the number of decoding belief propagation iterations in the Wiberg decoding tree. Overall, this work has the potential to advance multiple fields, from Information Theory, DNN architecture design (sparse and structured prior graph topology), efficient hardware design for Quantum and Classical DPU/TPU (graph, quantize and shift register architect.) to Materials Science and beyond.Comment: 71 pages, 42 Figures, 1 Table, 1 Appendix. arXiv admin note: text overlap with arXiv:2109.08184 by other author

    Theory and Practice of Computing with Excitable Dynamics

    Get PDF
    Reservoir computing (RC) is a promising paradigm for time series processing. In this paradigm, the desired output is computed by combining measurements of an excitable system that responds to time-dependent exogenous stimuli. The excitable system is called a reservoir and measurements of its state are combined using a readout layer to produce a target output. The power of RC is attributed to an emergent short-term memory in dynamical systems and has been analyzed mathematically for both linear and nonlinear dynamical systems. The theory of RC treats only the macroscopic properties of the reservoir, without reference to the underlying medium it is made of. As a result, RC is particularly attractive for building computational devices using emerging technologies whose structure is not exactly controllable, such as self-assembled nanoscale circuits. RC has lacked a formal framework for performance analysis and prediction that goes beyond memory properties. To provide such a framework, here a mathematical theory of memory and information processing in ordered and disordered linear dynamical systems is developed. This theory analyzes the optimal readout layer for a given task. The focus of the theory is a standard model of RC, the echo state network (ESN). An ESN consists of a fixed recurrent neural network that is driven by an external signal. The dynamics of the network is then combined linearly with readout weights to produce the desired output. The readout weights are calculated using linear regression. Using an analysis of regression equations, the readout weights can be calculated using only the statistical properties of the reservoir dynamics, the input signal, and the desired output. The readout layer weights can be calculated from a priori knowledge of the desired function to be computed and the weight matrix of the reservoir. This formulation explicitly depends on the input weights, the reservoir weights, and the statistics of the target function. This formulation is used to bound the expected error of the system for a given target function. The effects of input-output correlation and complex network structure in the reservoir on the computational performance of the system have been mathematically characterized. Far from the chaotic regime, ordered linear networks exhibit a homogeneous decay of memory in different dimensions, which keeps the input history coherent. As disorder is introduced in the structure of the network, memory decay becomes inhomogeneous along different dimensions causing decoherence in the input history, and degradation in task-solving performance. Close to the chaotic regime, the ordered systems show loss of temporal information in the input history, and therefore inability to solve tasks. However, by introducing disorder and therefore heterogeneous decay of memory the temporal information of input history is preserved and the task-solving performance is recovered. Thus for systems at the edge of chaos, disordered structure may enhance temporal information processing. Although the current framework only applies to linear systems, in principle it can be used to describe the properties of physical reservoir computing, e.g., photonic RC using short coherence-length light

    Consciousness as a State of Matter

    Full text link
    We examine the hypothesis that consciousness can be understood as a state of matter, "perceptronium", with distinctive information processing abilities. We explore five basic principles that may distinguish conscious matter from other physical systems such as solids, liquids and gases: the information, integration, independence, dynamics and utility principles. If such principles can identify conscious entities, then they can help solve the quantum factorization problem: why do conscious observers like us perceive the particular Hilbert space factorization corresponding to classical space (rather than Fourier space, say), and more generally, why do we perceive the world around us as a dynamic hierarchy of objects that are strongly integrated and relatively independent? Tensor factorization of matrices is found to play a central role, and our technical results include a theorem about Hamiltonian separability (defined using Hilbert-Schmidt superoperators) being maximized in the energy eigenbasis. Our approach generalizes Giulio Tononi's integrated information framework for neural-network-based consciousness to arbitrary quantum systems, and we find interesting links to error-correcting codes, condensed matter criticality, and the Quantum Darwinism program, as well as an interesting connection between the emergence of consciousness and the emergence of time.Comment: Replaced to match accepted CSF version; discussion improved, typos corrected. 36 pages, 15 fig

    Graph Clustering by Flow Simulation

    Get PDF

    Synchrony and bifurcations in coupled dynamical systems and effects of time delay

    Get PDF
    Dynamik auf Netzwerken ist ein mathematisches Feld, das in den letzten Jahrzehnten schnell gewachsen ist und Anwendungen in zahlreichen Disziplinen wie z.B. Physik, Biologie und Soziologie findet. Die Funktion vieler Netzwerke hĂ€ngt von der FĂ€higkeit ab, die Elemente des Netzwerkes zu synchronisieren. Mit anderen Worten, die Existenz und die transversale StabilitĂ€t der synchronen Mannigfaltigkeit sind zentrale Eigenschaften. Erst seit einigen Jahren wird versucht, den verwickelten Zusammenhang zwischen der Kopplungsstruktur und den StabilitĂ€tseigenschaften synchroner ZustĂ€nde zu verstehen. Genau das ist das zentrale Thema dieser Arbeit. ZunĂ€chst prĂ€sentiere ich erste Ergebnisse zur Klassifizierung der Kanten eines gerichteten Netzwerks bezĂŒglich ihrer Bedeutung fĂŒr die StabilitĂ€t des synchronen Zustands. Folgend untersuche ich ein komplexes Verzweigungsszenario in einem gerichteten Ring von Stuart-Landau Oszillatoren und zeige, dass das Szenario persistent ist, wenn dem Netzwerk eine schwach gewichtete Kante hinzugefĂŒgt wird. Daraufhin untersuche ich synchrone ZustĂ€nde in Ringen von Phasenoszillatoren die mit Zeitverzögerung gekoppelt sind. Ich bespreche die Koexistenz synchroner Lösungen und analysiere deren StabilitĂ€t und Verzweigungen. Weiter zeige ich, dass eine Zeitverschiebung genutzt werden kann, um Muster im Ring zu speichern und wiederzuerkennen. Diese Zeitverschiebung untersuche ich daraufhin fĂŒr beliebige Kopplungsstrukturen. Ich zeige, dass invariante Mannigfaltigkeiten des Flusses sowie ihre StabilitĂ€t unter der Zeitverschiebung erhalten bleiben. DarĂŒber hinaus bestimme ich die minimale Anzahl von Zeitverzögerungen, die gebraucht werden, um das System Ă€quivalent zu beschreiben. Schließlich untersuche ich das auffĂ€llige PhĂ€nomen eines nichtstetigen Übergangs zu SynchronizitĂ€t in Klassen großer Zufallsnetzwerke indem ich einen kĂŒrzlich eingefĂŒhrten Zugang zur Beschreibung großer Zufallsnetzwerke auf den Fall zeitverzögerter Kopplungen verallgemeinere.Since a couple of decades, dynamics on networks is a rapidly growing branch of mathematics with applications in various disciplines such as physics, biology or sociology. The functioning of many networks heavily relies on the ability to synchronize the network’s nodes. More precisely, the existence and the transverse stability of the synchronous manifold are essential properties. It was only in the last few years that people tried to understand the entangled relation between the coupling structure of a network, given by a (di-)graph, and the stability properties of synchronous states. This is the central theme of this dissertation. I first present results towards a classification of the links in a directed, diffusive network according to their impact on the stability of synchronization. Then I investigate a complex bifurcation scenario observed in a directed ring of Stuart-Landau oscillators. I show that under the addition of a single weak link, this scenario is persistent. Subsequently, I investigate synchronous patterns in a directed ring of phase oscillators coupled with time delay. I discuss the coexistence of multiple of synchronous solutions and investigate their stability and bifurcations. I apply these results by showing that a certain time-shift transformation can be used in order to employ the ring as a pattern recognition device. Next, I investigate the same time-shift transformation for arbitrary coupling structures in a very general setting. I show that invariant manifolds of the flow together with their stability properties are conserved under the time-shift transformation. Furthermore, I determine the minimal number of delays needed to equivalently describe the system’s dynamics. Finally, I investigate a peculiar phenomenon of non-continuous transition to synchrony observed in certain classes of large random networks, generalizing a recently introduced approach for the description of large random networks to the case of delayed couplings

    Continuous attractor working memory and provenance of channel models

    Get PDF
    The brain is a complex biological system composed of a multitude of microscopic processes, which together give rise to computational abilities observed in everyday behavior. Neuronal modeling, consisting of models of single neurons and neuronal networks at varying levels of biological detail, can synthesize the gaps currently hard to constrain in experiments and provide mechanistic explanations of how these computations might arise. In this thesis, I present two parallel lines of research on neuronal modeling, situated at varying levels of biological detail. First, I assess the provenance of voltage-gated ion channel models in an integrative meta-analysis that investigates a backlog of nearly 50 years of published research. To cope with the ever-increasing volume of research produced in the field of neuroscience, we need to develop methods for the systematic assessment and comparison of published work. As we demonstrate, neuronal models offer the intriguing possibility of performing automated quantitative analyses across studies, by standardized simulated experiments. We developed protocols for the quantitative comparison of voltage-gated ion channels, and applied them to a large body of published models, allowing us to assess the variety and temporal development of different models for the same ion channels over the time scale of years of research. Beyond a systematic classification of the existing body of research made available in an online platform, we show that our approach extends to large-scale comparisons of ion channel models to experimental data, thereby facilitating field-wide standardization of experimentally-constrained modeling. Second, I investigate neuronal models of working memory (WM). How can cortical networks bridge the short time scales of their microscopic components, which operate on the order of milliseconds, to the behaviorally relevant time scales of seconds observed in WM experiments? I consider here a candidate model: continuous attractor networks. These can implement WM for a continuum of possible spatial locations over several seconds and have been proposed for the organization of prefrontal cortical networks. I first present a novel method for the efficient prediction of the network-wide steady states from the underlying microscopic network properties. The method can be applied to predict and tune the "bump" shapes of continuous attractors implemented in networks of spiking neuron models connected by nonlinear synapses, which we demonstrate for saturating synapses involving NMDA receptors. In a second part, I investigate the computational role of short-term synaptic plasticity as a synaptic nonlinearity. Continuous attractor models are sensitive to the inevitable variability of biological neurons: variable neuronal firing and heterogeneous networks decrease the time that memories are accurately retained, eventually leading to a loss of memory functionality on behaviorally relevant time scales. In theory and simulations, I show that short-term plasticity can control the time scale of memory retention, with facilitation and depression playing antagonistic roles in controlling the drift and diffusion of locations in memory. Finally, we place quantitative constraints on the combination of synaptic and network parameters under which continuous attractors networks can implement reliable WM in cortical settings

    Evoked Patterns of Oscillatory Activity in Mean-Field Neuronal Networks

    Get PDF
    Oscillatory behaviors in populations of neurons are oberved in diverse contexts. In tasks involving working memory, a form of short-term memory, oscillations in different frequency bands have been shown to increase across varying spatial scales using recording methods such as EEG (electroencephalogram) and MEG (magnetoencephalogram). Such oscillatory activity has also been observed in the context of neural binding, where different features of objects that are perceived or recalled are associated with one another. These sets of data suggest that oscillatory dynamics may also play a key role in the maintenance and manipulation of items in working memory. Using similar recording techniques, including EEG and MEG, oscillatory neuronal activity has also been seen to occur when certain images that cause aversion and headaches in healthy human subjects or seizures in those with pattern-sensitive epilepsy are presented. The images most likely to cause such responses are those with dominant spatial frequencies near 3--5 cycles per degree, the same band of wavenumbers to which normal human vision exhibits the greatest contrast sensitivity. We model these oscillatory behaviors using mean-field, Wilson-Cowan-type neuronal networks. In the case of working memory and binding, we find that including the activity of certain long-lasting excitatory synapses in addition to the usual inhibitory and shorter-term excitatory synaptic activity allows for bistability between a low steady state and a high oscillatory state. By coupling several such populations together, both in-phase and out-of-phase oscillations arise, corresponding to distinct and bound items in working memory, respectively. We analyze the network's dynamics and dependence on biophysically relevant parameters using a combination of techniques, including numerical bifurcation analysis and weak coupling theory. In the case of spatially resonant responses to static simtuli, we employ Wilson-Cowan networks extended in one and two spatial dimensions. By placing the networks near Turing-Hopf bifurcations, we find they exhibit spatial resonances that compare well with empirical results. Using simulations, numerical bifurcation analysis, and perturbation theory, we characterize the observed dynamics and gain mathematical insight into the mechanisms that lead to these dynamics

    Modelling the neuromechanics of exploration and taxis in larval Drosophiila

    Get PDF
    The Drosophila larva is emerging as a useful tool in the study of complex behaviours, due to its relatively small size, its genetic tractability, and its varied behavioural repertoire. The larva executes a stereotypical exploratory routine that appears to consist of stochastic alternation between straight peristaltic crawling and reorientation events through lateral bending. The larva performs taxis by biasing this behavioural pattern, allowing it to move up or down attractive and aversive stimulus gradients. Existing explanations of exploration and taxis behaviour often neglect the larva's embodiment, focusing on central pattern generation and decision making circuits within the nervous system. In Chapter 1 of this thesis, I review the current state of knowledge regarding larval peristalsis, exploration, and taxis behaviours, as well as existing theories of their generation. I argue that an understanding of the animal's embodiment should lead to a deeper understanding of its behaviour. In Chapter 2, I present a model of the axial mechanics of the larva, and demonstrate how the animal's body physics can be exploited to produce peristalsis by using segmentally localised, positive feedback of strain rate. The mechanical model includes viscoelastic tissue mechanics, muscular inputs, and substrate interaction while sensory feedback is modelled as a linear feedback control law. In Chapter 3, I extend the mechanical model to study motion in the plane, including both axial and transverse deformations of the body. The feedback law is replaced by a simple model of the larval nervous system. The model includes both a segmentally localised reflex arc as well as long-range, mutual inhibition between segments. The complete model is capable of generating both peristalsis and spontaneous reorientation, leading to emergent exploration behaviour in the form of a deterministic superdiffusion process grounded in the chaotic mechanics of the larva's body. In Chapter 4, I consider taxis behaviour. I introduce a transverse reflex capable of modulating the effective transverse viscosity of the larval body. When the larva is experiencing an increasing attractive (aversive) stimulus, the reflex acts to increase (decrease) the effective transverse viscosity, causing bending to occur less (more) easily. As a result, the model larvae approach attractive stimuli and avoid aversive stumuli. On a population level, I show that the transverse reflex can be thought of as biasing the model animals towards sub- or super-diffusion. I compare the statistics of this behaviour to those of the real larva. In Chapter 5, I shift focus to engineered soft systems. Having successfully deployed an energy-based modelling approach in Chapters 2--4, I argue for the adoption of an energy-focused (specifically, port-Hamiltonian) approach within the field of soft robotics. In Chapter 6, I present some initial theoretical extensions to the models presented in chapter 2--4. I first focus on the mechanics of self-righting and rolling behaviours, before modelling the ventral nerve cord of the larva using a ring attractor architecture. Finally, in Chapter 7, I summarise the results of the previous chapters and discuss directions for future research

    Graph clustering by flow simulation

    Get PDF
    Dit proefschrift heeft als onderwerp het clusteren van grafen door middel van simulatie van stroming, een probleem dat in zijn algemeenheid behoort tot het gebied der clusteranalyse. In deze tak van wetenschap ontwerpt en onderzoekt men methoden die gegeven bepaalde data een onderverdeling in groepen genereren, waarbij het oogmerk is een onderverdeling in groepen te vinden die natuurlijk is. Dat wil zeggen dat verschillende data-elementen in dezelfde groep idealiter veel op elkaar lijken, en dat data-elementen uit verschillende groepen idealiter veel van elkaar verschillen. Soms ontbreken zulke groepjes helemaal; dan is er weinig patroon te herkennen in de data. Het idee is dat de aanwezigheid van natuurlijke groepjes het mogelijk maakt de data te categoriseren. Een voorbeeld is het clusteren van gegevens (over symptomen of lichaamskarakteristieken) van patienten die aan dezelfde ziekte lijden. Als er duidelijke groepjes bestaan in die gegevens, kan dit tot extra inzicht leiden in de ziekte. Clusteranalyse kan aldus gebruikt worden voor exploratief onderzoek. Verdere voorbeelden komen uit de scheikunde, taxonomie, psychiatrie, archeologie, marktonderzoek en nog vele andere disicplines. Taxonomie, de studie van de classificatie van organismen, heeft een rijke geschiedenis beginnend bij Aristoteles en culminerend in de werken van Linnaeus. In feite kan de clusteranalyse gezien worden als het resultaat van een steeds meer systematische en abstracte studie van de diverse methoden ontworpen in verschillende toepassingsgebieden, waarbij methode zowel wordt gescheiden van data en toepassingsgebied als van berekeningswijze. In de cluster analyse kunnen grofweg twee richtingen onderscheiden worden, naar gelang het type data dat geclassificeerd moet worden. De data-elementen in het voorbeeld hierboven worden beschreven door vectoren (lijstjes van scores of metingen), en het verschil tussen twee elementen wordt bepaald door het verschil van de vectoren. Deze dissertatie betreft cluster analyse toegepast op data van het type `graaf'. Voorbeelden komen uit de patroonherkenning, het computer ondersteund ontwerpen, databases voorzien van hyperlinks en het World Wide Web. In al deze gevallen is er sprake van `punten' die verbonden zijn of niet. Een stelsel van punten samen met hun verbindingen heet een graaf. Een goede clustering van een graaf deelt de punten op in groepjes zodanig dat er weinig verbindingen lopen tussen (punten uit) verschillende groepjes en er veel verbindingen zijn in elk groepje afzonderlijk
    corecore