2,964 research outputs found
Spiking Neural P Systems with Addition/Subtraction Computing on Synapses
Spiking neural P systems (SN P systems, for short) are a class of distributed
and parallel computing models inspired from biological spiking neurons. In this paper,
we introduce a variant called SN P systems with addition/subtraction computing on
synapses (CSSN P systems). CSSN P systems are inspired and motivated by the shunting
inhibition of biological synapses, while incorporating ideas from dynamic graphs and
networks. We consider addition and subtraction operations on synapses, and prove that
CSSN P systems are computationally universal as number generators, under a normal
form (i.e. a simplifying set of restrictions)
Neuromorphic Hardware In The Loop: Training a Deep Spiking Network on the BrainScaleS Wafer-Scale System
Emulating spiking neural networks on analog neuromorphic hardware offers
several advantages over simulating them on conventional computers, particularly
in terms of speed and energy consumption. However, this usually comes at the
cost of reduced control over the dynamics of the emulated networks. In this
paper, we demonstrate how iterative training of a hardware-emulated network can
compensate for anomalies induced by the analog substrate. We first convert a
deep neural network trained in software to a spiking network on the BrainScaleS
wafer-scale neuromorphic system, thereby enabling an acceleration factor of 10
000 compared to the biological time domain. This mapping is followed by the
in-the-loop training, where in each training step, the network activity is
first recorded in hardware and then used to compute the parameter updates in
software via backpropagation. An essential finding is that the parameter
updates do not have to be precise, but only need to approximately follow the
correct gradient, which simplifies the computation of updates. Using this
approach, after only several tens of iterations, the spiking network shows an
accuracy close to the ideal software-emulated prototype. The presented
techniques show that deep spiking networks emulated on analog neuromorphic
devices can attain good computational performance despite the inherent
variations of the analog substrate.Comment: 8 pages, 10 figures, submitted to IJCNN 201
- …