561 research outputs found
What grid cells convey about rat location
We characterize the relationship between the simultaneously recorded quantities of rodent grid cell firing and the position of the rat. The formalization reveals various properties of grid cell activity when considered as a neural code for representing and updating estimates of the rat's location. We show that, although the spatially periodic response of grid cells appears wasteful, the code is fully combinatorial in capacity. The resulting range for unambiguous position representation is vastly greater than the β1β10 m periods of individual lattices, allowing for unique high-resolution position specification over the behavioral foraging ranges of rats, with excess capacity that could be used for error correction. Next, we show that the merits of the grid cell code for position representation extend well beyond capacity and include arithmetic properties that facilitate position updating. We conclude by considering the numerous implications, for downstream readouts and experimental tests, of the properties of the grid cell code
Gradient-trained Weights in Wide Neural Networks Align Layerwise to Error-scaled Input Correlations
Recent works have examined how deep neural networks, which can solve a
variety of difficult problems, incorporate the statistics of training data to
achieve their success. However, existing results have been established only in
limited settings. In this work, we derive the layerwise weight dynamics of
infinite-width neural networks with nonlinear activations trained by gradient
descent. We show theoretically that weight updates are aligned with input
correlations from intermediate layers weighted by error, and demonstrate
empirically that the result also holds in finite-width wide networks. The
alignment result allows us to formulate backpropagation-free learning rules,
named Align-zero and Align-ada, that theoretically achieve the same alignment
as backpropagation. Finally, we test these learning rules on benchmark problems
in feedforward and recurrent neural networks and demonstrate, in wide networks,
comparable performance to backpropagation.Comment: 22 pages, 11 figure
Testing odor response stereotypy in the Drosophila mushroom body
The mushroom body is an insect brain structure required for olfactory learning. Its principal neurons, the Kenyon cells (KCs), form a large cell population. The neuronal populations from which their olfactory input derives (olfactory sensory and projection neurons) can be identified individually by genetic, anatomical, and physiological criteria. We ask whether KCs are similarly identifiable individually, using genetic markers and whole-cell patch-clamp in vivo. We find that across-animal responses are as diverse within the genetically labeled subset as across all KCs in a larger sample. These results combined with those from a simple model, using projection neuron odor responses as inputs, suggest that the precise circuit specification seen at earlier stages of odor processing is likely absent among the mushroom body KCs
Beyond Geometry: Comparing the Temporal Structure of Computation in Neural Circuits with Dynamical Similarity Analysis
How can we tell whether two neural networks are utilizing the same internal
processes for a particular computation? This question is pertinent for multiple
subfields of both neuroscience and machine learning, including neuroAI,
mechanistic interpretability, and brain-machine interfaces. Standard approaches
for comparing neural networks focus on the spatial geometry of latent states.
Yet in recurrent networks, computations are implemented at the level of neural
dynamics, which do not have a simple one-to-one mapping with geometry. To
bridge this gap, we introduce a novel similarity metric that compares two
systems at the level of their dynamics. Our method incorporates two components:
Using recent advances in data-driven dynamical systems theory, we learn a
high-dimensional linear system that accurately captures core features of the
original nonlinear dynamics. Next, we compare these linear approximations via a
novel extension of Procrustes Analysis that accounts for how vector fields
change under orthogonal transformation. Via four case studies, we demonstrate
that our method effectively identifies and distinguishes dynamic structure in
recurrent neural networks (RNNs), whereas geometric methods fall short. We
additionally show that our method can distinguish learning rules in an
unsupervised manner. Our method therefore opens the door to novel data-driven
analyses of the temporal structure of neural computation, and to more rigorous
testing of RNNs as models of the brain.Comment: 21 pages, 10 figure
- β¦