1,209 research outputs found

    Pattern formation in oscillatory complex networks consisting of excitable nodes

    Full text link
    Oscillatory dynamics of complex networks has recently attracted great attention. In this paper we study pattern formation in oscillatory complex networks consisting of excitable nodes. We find that there exist a few center nodes and small skeletons for most oscillations. Complicated and seemingly random oscillatory patterns can be viewed as well-organized target waves propagating from center nodes along the shortest paths, and the shortest loops passing through both the center nodes and their driver nodes play the role of oscillation sources. Analyzing simple skeletons we are able to understand and predict various essential properties of the oscillations and effectively modulate the oscillations. These methods and results will give insights into pattern formation in complex networks, and provide suggestive ideas for studying and controlling oscillations in neural networks.Comment: 15 pages, 7 figures, to appear in Phys. Rev.

    Noise Induced Coherence in Neural Networks

    Full text link
    We investigate numerically the dynamics of large networks of NN globally pulse-coupled integrate and fire neurons in a noise-induced synchronized state. The powerspectrum of an individual element within the network is shown to exhibit in the thermodynamic limit (N→∞N\to \infty) a broadband peak and an additional delta-function peak that is absent from the powerspectrum of an isolated element. The powerspectrum of the mean output signal only exhibits the delta-function peak. These results are explained analytically in an exactly soluble oscillator model with global phase coupling.Comment: 4 pages ReVTeX and 3 postscript figure

    Event-driven simulations of a plastic, spiking neural network

    Full text link
    We consider a fully-connected network of leaky integrate-and-fire neurons with spike-timing-dependent plasticity. The plasticity is controlled by a parameter representing the expected weight of a synapse between neurons that are firing randomly with the same mean frequency. For low values of the plasticity parameter, the activities of the system are dominated by noise, while large values of the plasticity parameter lead to self-sustaining activity in the network. We perform event-driven simulations on finite-size networks with up to 128 neurons to find the stationary synaptic weight conformations for different values of the plasticity parameter. In both the low and high activity regimes, the synaptic weights are narrowly distributed around the plasticity parameter value consistent with the predictions of mean-field theory. However, the distribution broadens in the transition region between the two regimes, representing emergent network structures. Using a pseudophysical approach for visualization, we show that the emergent structures are of "path" or "hub" type, observed at different values of the plasticity parameter in the transition region.Comment: 9 pages, 6 figure

    Dynamic range in the C.elegans brain network

    Get PDF
    We study external electrical perturbations and their responses in the brain dynamic network of the Caenorhabditis eleganssoil worm, given by the connectome of its large somatic nervous system. Our analysis is inspired by a realistic experiment where one stimulates externally specific parts of the brain and studies the persistent neural activity triggered in other cortical regions. In this work, we perturb groups of neurons that form communities, identified by the walktrap community detection method, by trains of stereotypical electrical Poissonian impulses and study the propagation of neural activity to other communities by measuring the corresponding dynamic ranges and Steven law exponents. We show that when one perturbs specific communities, keeping the rest unperturbed, the external stimulations are able to propagate to some of them but not to all. There are also perturbations that do not trigger any response. We found that this depends on the initially perturbed community. Finally, we relate our findings for the former cases with low neural synchronization, self-criticality, and large information flow capacity, and interpret them as the ability of the brainnetwork to respond to external perturbations when it works at criticality and its information flow capacity becomes maximal

    Supervised Learning in Multilayer Spiking Neural Networks

    Get PDF
    The current article introduces a supervised learning algorithm for multilayer spiking neural networks. The algorithm presented here overcomes some limitations of existing learning algorithms as it can be applied to neurons firing multiple spikes and it can in principle be applied to any linearisable neuron model. The algorithm is applied successfully to various benchmarks, such as the XOR problem and the Iris data set, as well as complex classifications problems. The simulations also show the flexibility of this supervised learning algorithm which permits different encodings of the spike timing patterns, including precise spike trains encoding.Comment: 38 pages, 4 figure

    Triggering up states in all-to-all coupled neurons

    Full text link
    Slow-wave sleep in mammalians is characterized by a change of large-scale cortical activity currently paraphrased as cortical Up/Down states. A recent experiment demonstrated a bistable collective behaviour in ferret slices, with the remarkable property that the Up states can be switched on and off with pulses, or excitations, of same polarity; whereby the effect of the second pulse significantly depends on the time interval between the pulses. Here we present a simple time discrete model of a neural network that exhibits this type of behaviour, as well as quantitatively reproduces the time-dependence found in the experiments.Comment: epl Europhysics Letters, accepted (2010

    Competing synapses with two timescales: a basis for learning and forgetting

    Full text link
    Competitive dynamics are thought to occur in many processes of learning involving synaptic plasticity. Here we show, in a game theory-inspired model of synaptic interactions, that the competition between synapses in their weak and strong states gives rise to a natural framework of learning, with the prediction of memory inherent in a timescale for `forgetting' a learned signal. Among our main results is the prediction that memory is optimized if the weak synapses are really weak, and the strong synapses are really strong. Our work admits of many extensions and possible experiments to test its validity, and in particular might complement an existing model of reaching, which has strong experimental support.Comment: 7 pages, 3 figures, to appear in Europhysics Letter

    Adherent carbon film deposition by cathodic arc with implantation

    Get PDF
    A method of improving the adhesion of carbon thin films deposited using a cathodic vacuum arc by the use of implantation at energies up to 20 keV is described. A detailed analysis of carbon films deposited onto silicon in this way is carried out using complementary techniques of transmission electron microscopy and x-ray photoelectron spectroscopy (XPS) is presented. This analysis shows that an amorphous mixing layer consisting of carbon and silicon is formed between the grown pure carbon film and the crystalline silicon substrate. In the mixing layer, it is shown that some chemical bonding occurs between carbon and silicon. Damage to the underlying crystalline silicon substrate is observed and believed to be caused by interstitial implanted carbon atoms which XPS shows are not bonded to the silicon. The effectiveness of this technique is confirmed by scratch testing and by analysis with scanning electron microscopy which shows failure of the silicon substrate occurs before delamination of the carbon film
    • …
    corecore