889 research outputs found

    Adaptive regulation of sparseness by feedforward inhibition

    Get PDF
    In the mushroom body of insects, odors are represented by very few spikes in a small number of neurons, a highly efficient strategy known as sparse coding. Physiological studies of these neurons have shown that sparseness is maintained across thousand-fold changes in odor concentration. Using a realistic computational model, we propose that sparseness in the olfactory system is regulated by adaptive feedforward inhibition. When odor concentration changes, feedforward inhibition modulates the duration of the temporal window over which the mushroom body neurons may integrate excitatory presynaptic input. This simple adaptive mechanism could maintain the sparseness of sensory representations across wide ranges of stimulus conditions

    Maximum Performance at Minimum Cost in Network Synchronization

    Full text link
    We consider two optimization problems on synchronization of oscillator networks: maximization of synchronizability and minimization of synchronization cost. We first develop an extension of the well-known master stability framework to the case of non-diagonalizable Laplacian matrices. We then show that the solution sets of the two optimization problems coincide and are simultaneously characterized by a simple condition on the Laplacian eigenvalues. Among the optimal networks, we identify a subclass of hierarchical networks, characterized by the absence of feedback loops and the normalization of inputs. We show that most optimal networks are directed and non-diagonalizable, necessitating the extension of the framework. We also show how oriented spanning trees can be used to explicitly and systematically construct optimal networks under network topological constraints. Our results may provide insights into the evolutionary origin of structures in complex networks for which synchronization plays a significant role.Comment: 29 pages, 9 figures, accepted for publication in Physica D, minor correction

    Synchronized dynamics of cortical neurons with time-delay feedback

    Get PDF
    The dynamics of three mutually coupled cortical neurons with time delays in the coupling are explored numerically and analytically. The neurons are coupled in a line, with the middle neuron sending a somewhat stronger projection to the outer neurons than the feedback it receives, to model for instance the relay of a signal from primary to higher cortical areas. For a given coupling architecture, the delays introduce correlations in the time series at the time-scale of the delay. It was found that the middle neuron leads the outer ones by the delay time, while the outer neurons are synchronized with zero lag times. Synchronization is found to be highly dependent on the synaptic time constant, with faster synapses increasing both the degree of synchronization and the firing rate. Analysis shows that presynaptic input during the interspike interval stabilizes the synchronous state, even for arbitrarily weak coupling, and independent of the initial phase. The finding may be of significance to synchronization of large groups of cells in the cortex that are spatially distanced from each other.Comment: 21 pages, 11 figure

    Information transmission in oscillatory neural activity

    Full text link
    Periodic neural activity not locked to the stimulus or to motor responses is usually ignored. Here, we present new tools for modeling and quantifying the information transmission based on periodic neural activity that occurs with quasi-random phase relative to the stimulus. We propose a model to reproduce characteristic features of oscillatory spike trains, such as histograms of inter-spike intervals and phase locking of spikes to an oscillatory influence. The proposed model is based on an inhomogeneous Gamma process governed by a density function that is a product of the usual stimulus-dependent rate and a quasi-periodic function. Further, we present an analysis method generalizing the direct method (Rieke et al, 1999; Brenner et al, 2000) to assess the information content in such data. We demonstrate these tools on recordings from relay cells in the lateral geniculate nucleus of the cat.Comment: 18 pages, 8 figures, to appear in Biological Cybernetic

    Self-organization in the olfactory system: one shot odor recognition in insects

    Get PDF
    We show in a model of spiking neurons that synaptic plasticity in the mushroom bodies in combination with the general fan-in, fan-out properties of the early processing layers of the olfactory system might be sufficient to account for its efficient recognition of odors. For a large variety of initial conditions the model system consistently finds a working solution without any fine-tuning, and is, therefore, inherently robust. We demonstrate that gain control through the known feedforward inhibition of lateral horn interneurons increases the capacity of the system but is not essential for its general function. We also predict an upper limit for the number of odor classes Drosophila can discriminate based on the number and connectivity of its olfactory neurons

    Gain control network conditions in early sensory coding

    Get PDF
    Gain control is essential for the proper function of any sensory system. However, the precise mechanisms for achieving effective gain control in the brain are unknown. Based on our understanding of the existence and strength of connections in the insect olfactory system, we analyze the conditions that lead to controlled gain in a randomly connected network of excitatory and inhibitory neurons. We consider two scenarios for the variation of input into the system. In the first case, the intensity of the sensory input controls the input currents to a fixed proportion of neurons of the excitatory and inhibitory populations. In the second case, increasing intensity of the sensory stimulus will both, recruit an increasing number of neurons that receive input and change the input current that they receive. Using a mean field approximation for the network activity we derive relationships between the parameters of the network that ensure that the overall level of activity of the excitatory population remains unchanged for increasing intensity of the external stimulation. We find that, first, the main parameters that regulate network gain are the probabilities of connections from the inhibitory population to the excitatory population and of the connections within the inhibitory population. Second, we show that strict gain control is not achievable in a random network in the second case, when the input recruits an increasing number of neurons. Finally, we confirm that the gain control conditions derived from the mean field approximation are valid in simulations of firing rate models and Hodgkin-Huxley conductance based models

    Stimulus-dependent maximum entropy models of neural population codes

    Get PDF
    Neural populations encode information about their stimulus in a collective fashion, by joint activity patterns of spiking and silence. A full account of this mapping from stimulus to neural activity is given by the conditional probability distribution over neural codewords given the sensory input. To be able to infer a model for this distribution from large-scale neural recordings, we introduce a stimulus-dependent maximum entropy (SDME) model---a minimal extension of the canonical linear-nonlinear model of a single neuron, to a pairwise-coupled neural population. The model is able to capture the single-cell response properties as well as the correlations in neural spiking due to shared stimulus and due to effective neuron-to-neuron connections. Here we show that in a population of 100 retinal ganglion cells in the salamander retina responding to temporal white-noise stimuli, dependencies between cells play an important encoding role. As a result, the SDME model gives a more accurate account of single cell responses and in particular outperforms uncoupled models in reproducing the distributions of codewords emitted in response to a stimulus. We show how the SDME model, in conjunction with static maximum entropy models of population vocabulary, can be used to estimate information-theoretic quantities like surprise and information transmission in a neural population.Comment: 11 pages, 7 figure

    Competition-based model of pheromone component ratio detection in the moth

    Get PDF
    For some moth species, especially those closely interrelated and sympatric, recognizing a specific pheromone component concentration ratio is essential for males to successfully locate conspecific females. We propose and determine the properties of a minimalist competition-based feed-forward neuronal model capable of detecting a certain ratio of pheromone components independently of overall concentration. This model represents an elementary recognition unit for the ratio of binary mixtures which we propose is entirely contained in the macroglomerular complex (MGC) of the male moth. A set of such units, along with projection neurons (PNs), can provide the input to higher brain centres. We found that (1) accuracy is mainly achieved by maintaining a certain ratio of connection strengths between olfactory receptor neurons (ORN) and local neurons (LN), much less by properties of the interconnections between the competing LNs proper. An exception to this rule is that it is beneficial if connections between generalist LNs (i.e. excited by either pheromone component) and specialist LNs (i.e. excited by one component only) have the same strength as the reciprocal specialist to generalist connections. (2) successful ratio recognition is achieved using latency-to-first-spike in the LN populations which, in contrast to expectations with a population rate code, leads to a broadening of responses for higher overall concentrations consistent with experimental observations. (3) when longer durations of the competition between LNs were observed it did not lead to higher recognition accuracy

    Crusading for a Christian Nation: How Christian Nationalist Pastors Have Disseminated Donald Trump's Myth of the Stolen Election

    Get PDF
    With the Capitol riot on January 6, 2021, being one of the latest manifestations of the deep crisis that American democracy is facing, this essay focuses on the fusion of religion and politics as manifested in the Christian nationalist movement – a belief system rooted in the superior morality of Christianity and the conviction that both American society and politics should be exclusively guided by Christian principles. Three sermons by altogether three pastors – namely Greg Locke, Ken Peters, and Brian Gibson – serve as the basis for this analysis, which focuses on the following question: How exactly did they interpret Trump’s loss in the 2020 presidential election, and how have they supported the narrative of a stolen election? The myth of a stolen election (the Big Lie) was introduced by former president Donald Trump himself after his loss in the 2020 election to 'stop the steal' and secure his second term in office by mobilizing his loyal followers. The analysis will prove that the pseudo-religious movement played a major role in consolidating this narrative by using biblical language and scripture, apocalyptic visions, conspiracy theories, as well as violent rhetoric to support their president. Eventually, it was exactly this politico-religious agenda and religious fervor that promoted violence and hatred, which facilitated the violent insurrection at the U.S. Capitol – an incident that has shaken American democracy to the core
    corecore