1,782 research outputs found
Arithmetic Operations with Spiking Neural P Systems with Rules and Weights on Synapses
The application of spiking neural P systems with rules and weights on synapses to arithmetic operations is discussed in this paper. We design specific spiking neural P systems with rules and weights on synapses for successfully performing addition, multiplication and the greatest common divisor. This is the first attempt to discuss the application of the new variant of spiking neural P systems, spiking neural P systems with rules and weights on synapses, and especially the use of spiking neural P systems to perform the greatest common divisor. Comparing with the results reported in the literature, smaller number of neurons are required to fulfill the arithmetic operations
Simulating Spiking Neural P systems without delays using GPUs
We present in this paper our work regarding simulating a type of P system
known as a spiking neural P system (SNP system) using graphics processing units
(GPUs). GPUs, because of their architectural optimization for parallel
computations, are well-suited for highly parallelizable problems. Due to the
advent of general purpose GPU computing in recent years, GPUs are not limited
to graphics and video processing alone, but include computationally intensive
scientific and mathematical applications as well. Moreover P systems, including
SNP systems, are inherently and maximally parallel computing models whose
inspirations are taken from the functioning and dynamics of a living cell. In
particular, SNP systems try to give a modest but formal representation of a
special type of cell known as the neuron and their interactions with one
another. The nature of SNP systems allowed their representation as matrices,
which is a crucial step in simulating them on highly parallel devices such as
GPUs. The highly parallel nature of SNP systems necessitate the use of hardware
intended for parallel computations. The simulation algorithms, design
considerations, and implementation are presented. Finally, simulation results,
observations, and analyses using an SNP system that generates all numbers in
- {1} are discussed, as well as recommendations for future work.Comment: 19 pages in total, 4 figures, listings/algorithms, submitted at the
9th Brainstorming Week in Membrane Computing, University of Seville, Spai
Stochastic rounding and reduced-precision fixed-point arithmetic for solving neural ordinary differential equations
Although double-precision floating-point arithmetic currently dominates
high-performance computing, there is increasing interest in smaller and simpler
arithmetic types. The main reasons are potential improvements in energy
efficiency and memory footprint and bandwidth. However, simply switching to
lower-precision types typically results in increased numerical errors. We
investigate approaches to improving the accuracy of reduced-precision
fixed-point arithmetic types, using examples in an important domain for
numerical computation in neuroscience: the solution of Ordinary Differential
Equations (ODEs). The Izhikevich neuron model is used to demonstrate that
rounding has an important role in producing accurate spike timings from
explicit ODE solution algorithms. In particular, fixed-point arithmetic with
stochastic rounding consistently results in smaller errors compared to single
precision floating-point and fixed-point arithmetic with round-to-nearest
across a range of neuron behaviours and ODE solvers. A computationally much
cheaper alternative is also investigated, inspired by the concept of dither
that is a widely understood mechanism for providing resolution below the least
significant bit (LSB) in digital signal processing. These results will have
implications for the solution of ODEs in other subject areas, and should also
be directly relevant to the huge range of practical problems that are
represented by Partial Differential Equations (PDEs).Comment: Submitted to Philosophical Transactions of the Royal Society
Spiking Neural P Systems with Addition/Subtraction Computing on Synapses
Spiking neural P systems (SN P systems, for short) are a class of distributed
and parallel computing models inspired from biological spiking neurons. In this paper,
we introduce a variant called SN P systems with addition/subtraction computing on
synapses (CSSN P systems). CSSN P systems are inspired and motivated by the shunting
inhibition of biological synapses, while incorporating ideas from dynamic graphs and
networks. We consider addition and subtraction operations on synapses, and prove that
CSSN P systems are computationally universal as number generators, under a normal
form (i.e. a simplifying set of restrictions)
Building Blocks for Spikes Signals Processing
Neuromorphic engineers study models and
implementations of systems that mimic neurons behavior in the
brain. Neuro-inspired systems commonly use spikes to
represent information. This representation has several
advantages: its robustness to noise thanks to repetition, its
continuous and analog information representation using digital
pulses, its capacity of pre-processing during transmission time,
... , Furthermore, spikes is an efficient way, found by nature, to
codify, transmit and process information. In this paper we
propose, design, and analyze neuro-inspired building blocks
that can perform spike-based analog filters used in signal
processing. We present a VHDL implementation for FPGA.
Presented building blocks take advantages of the spike rate
coded representation to perform a massively parallel processing
without complex hardware units, like floating point arithmetic
units, or a large memory. Those low requirements of hardware
allow the integration of a high number of blocks inside a FPGA,
allowing to process fully in parallel several spikes coded signals.Junta de Andalucía P06-TIC-O1417Ministerio de Ciencia e Innovación TEC2009-10639-C04-02Ministerio de Ciencia e Innovación TEC2006-11730-C03-0
Time After Time: Notes on Delays In Spiking Neural P Systems
Spiking Neural P systems, SNP systems for short, are biologically inspired
computing devices based on how neurons perform computations. SNP systems use
only one type of symbol, the spike, in the computations. Information is encoded
in the time differences of spikes or the multiplicity of spikes produced at
certain times. SNP systems with delays (associated with rules) and those
without delays are two of several Turing complete SNP system variants in
literature. In this work we investigate how restricted forms of SNP systems
with delays can be simulated by SNP systems without delays. We show the
simulations for the following spike routing constructs: sequential, iteration,
join, and split.Comment: 11 pages, 9 figures, 4 lemmas, 1 theorem, preprint of Workshop on
Computation: Theory and Practice 2012 at DLSU, Manila together with UP
Diliman, DLSU, Tokyo Institute of Technology, and Osaka universit
Neuromorphic, Digital and Quantum Computation with Memory Circuit Elements
Memory effects are ubiquitous in nature and the class of memory circuit
elements - which includes memristors, memcapacitors and meminductors - shows
great potential to understand and simulate the associated fundamental physical
processes. Here, we show that such elements can also be used in electronic
schemes mimicking biologically-inspired computer architectures, performing
digital logic and arithmetic operations, and can expand the capabilities of
certain quantum computation schemes. In particular, we will discuss few
examples where the concept of memory elements is relevant to the realization of
associative memory in neuronal circuits, spike-timing-dependent plasticity of
synapses, digital and field-programmable quantum computing
GeNN: a code generation framework for accelerated brain simulations
Large-scale numerical simulations of detailed brain circuit models are important for identifying hypotheses on brain functions and testing their consistency and plausibility. An ongoing challenge for simulating realistic models is, however, computational speed. In this paper, we present the GeNN (GPU-enhanced Neuronal Networks) framework, which aims to facilitate the use of graphics accelerators for computational models of large-scale neuronal networks to address this challenge. GeNN is an open source library that generates code to accelerate the execution of network simulations on NVIDIA GPUs, through a flexible and extensible interface, which does not require in-depth technical knowledge from the users. We present performance benchmarks showing that 200-fold speedup compared to a single core of a CPU can be achieved for a network of one million conductance based Hodgkin-Huxley neurons but that for other models the speedup can differ.
GeNN is available for Linux, Mac OS X and Windows platforms. The source code, user manual, tutorials,
Wiki, in-depth example projects and all other related information can be found on the project website http://genn-team.github.io/genn/
- …