58 research outputs found

    Threefold Flops via Matrix Factorization

    Get PDF
    The explicit McKay correspondence, as formulated by Gonzalez-Sprinberg and Verdier, associates to each exceptional divisor in the minimal resolution of a rational double point a matrix factorization of the equation of the rational double point. We study deformations of these matrix factorizations, and show that they exist over an appropriate "partially resolved" deformation space for rational double points of types A and D. As a consequence, all simple flops of lengths 1 and 2 can be described in terms of blowups defined from matrix factorizations. We also formulate conjectures which would extend these results to rational double points of type E and simple flops of length greater than 2.Comment: v2: minor change

    Flexible Memory Networks

    Get PDF
    Networks of neurons in some brain areas are flexible enough to encode new memories quickly. Using a standard firing rate model of recurrent networks, we develop a theory of flexible memory networks. Our main results characterize networks having the maximal number of flexible memory patterns, given a constraint graph on the network's connectivity matrix. Modulo a mild topological condition, we find a close connection between maximally flexible networks and rank 1 matrices. The topological condition is H_1(X;Z)=0, where X is the clique complex associated to the network's constraint graph; this condition is generically satisfied for large random networks that are not overly sparse. In order to prove our main results, we develop some matrix-theoretic tools and present them in a self-contained section independent of the neuroscience context.Comment: Accepted to Bulletin of Mathematical Biology, 11 July 201

    Matrix Model Superpotentials and Calabi-Yau Spaces: an ADE Classification

    Get PDF
    We use F. Ferrari's methods relating matrix models to Calabi-Yau spaces in order to explain Intriligator and Wecht's ADE classification of N=1\N=1 superconformal theories which arise as RG fixed points of N=1\N = 1 SQCD theories with adjoints. The connection between matrix models and N=1\N = 1 gauge theories can be seen as evidence for the Dijkgraaf--Vafa conjecture. We find that ADE superpotentials in the Intriligator--Wecht classification exactly match matrix model superpotentials obtained from Calabi-Yau's with corresponding ADE singularities. Moreover, in the additional \Hat{O}, \Hat{A}, \Hat{D} and \Hat{E} cases we find new singular geometries. These `hat' geometries are closely related to their ADE counterparts, but feature non-isolated singularities. As a byproduct, we give simple descriptions for small resolutions of Gorenstein threefold singularities in terms of transition functions between just two coordinate charts. To obtain these results we develop techniques for performing small resolutions and small blow-downs, including an algorithm for blowing down exceptional \PP^1's. In particular, we conjecture that small resolutions for isolated Gorenstein threefold singularities can be obtained by deforming matrix factorizations for simple surface singularities -- and prove this in the length 1 and length 2 cases.Comment: PhD thesis, 120 page

    Clique topology reveals intrinsic geometric structure in neural correlations

    Get PDF
    Detecting meaningful structure in neural activity and connectivity data is challenging in the presence of hidden nonlinearities, where traditional eigenvalue-based methods may be misleading. We introduce a novel approach to matrix analysis, called clique topology, that extracts features of the data invariant under nonlinear monotone transformations. These features can be used to detect both random and geometric structure, and depend only on the relative ordering of matrix entries. We then analyzed the activity of pyramidal neurons in rat hippocampus, recorded while the animal was exploring a two-dimensional environment, and confirmed that our method is able to detect geometric organization using only the intrinsic pattern of neural correlations. Remarkably, we found similar results during non-spatial behaviors such as wheel running and REM sleep. This suggests that the geometric structure of correlations is shaped by the underlying hippocampal circuits, and is not merely a consequence of position coding. We propose that clique topology is a powerful new tool for matrix analysis in biological settings, where the relationship of observed quantities to more meaningful variables is often nonlinear and unknown.Comment: 29 pages, 4 figures, 13 supplementary figures (last two authors contributed equally

    Diversity of emergent dynamics in competitive threshold-linear networks: a preliminary report

    Full text link
    Threshold-linear networks consist of simple units interacting in the presence of a threshold nonlinearity. Competitive threshold-linear networks have long been known to exhibit multistability, where the activity of the network settles into one of potentially many steady states. In this work, we find conditions that guarantee the absence of steady states, while maintaining bounded activity. These conditions lead us to define a combinatorial family of competitive threshold-linear networks, parametrized by a simple directed graph. By exploring this family, we discover that threshold-linear networks are capable of displaying a surprisingly rich variety of nonlinear dynamics, including limit cycles, quasiperiodic attractors, and chaos. In particular, several types of nonlinear behaviors can co-exist in the same network. Our mathematical results also enable us to engineer networks with multiple dynamic patterns. Taken together, these theoretical and computational findings suggest that threshold-linear networks may be a valuable tool for understanding the relationship between network connectivity and emergent dynamics.Comment: 12 pages, 9 figures. Preliminary repor

    Cell Groups Reveal Structure of Stimulus Space

    Get PDF
    An important task of the brain is to represent the outside world. It is unclear how the brain may do this, however, as it can only rely on neural responses and has no independent access to external stimuli in order to ‘‘decode’’ what those responses mean. We investigate what can be learned about a space of stimuli using only the action potentials (spikes) of cells with stereotyped—but unknown—receptive fields. Using hippocampal place cells as a model system, we show that one can (1) extract global features of the environment and (2) construct an accurate representation of space, up to an overall scale factor, that can be used to track the animal’s position. Unlike previous approaches to reconstructing position from place cell activity, this information is derived without knowing place fields or any other functions relating neural responses to position. We find that simply knowing which groups of cells fire together reveals a surprising amount of structure in the underlying stimulus space; this may enable the brain to construct its own internal representations

    Cell Groups Reveal Structure of Stimulus Space

    Get PDF
    An important task of the brain is to represent the outside world. It is unclear how the brain may do this, however, as it can only rely on neural responses and has no independent access to external stimuli in order to “decode” what those responses mean. We investigate what can be learned about a space of stimuli using only the action potentials (spikes) of cells with stereotyped—but unknown—receptive fields. Using hippocampal place cells as a model system, we show that one can (1) extract global features of the environment and (2) construct an accurate representation of space, up to an overall scale factor, that can be used to track the animal's position. Unlike previous approaches to reconstructing position from place cell activity, this information is derived without knowing place fields or any other functions relating neural responses to position. We find that simply knowing which groups of cells fire together reveals a surprising amount of structure in the underlying stimulus space; this may enable the brain to construct its own internal representations

    Hyperplane Neural Codes and the Polar Complex

    Full text link
    Hyperplane codes are a class of convex codes that arise as the output of a one layer feed-forward neural network. Here we establish several natural properties of stable hyperplane codes in terms of the {\it polar complex} of the code, a simplicial complex associated to any combinatorial code. We prove that the polar complex of a stable hyperplane code is shellable and show that most currently known properties of the hyperplane codes follow from the shellability of the appropriate polar complex.Comment: 23 pages, 5 figures. To appear in Proceedings of the Abel Symposiu
    corecore