140 research outputs found
Investigation of Synapto-dendritic Kernel Adapting Neuron models and their use in spiking neuromorphic architectures
The motivation for this thesis is idea that abstract, adaptive, hardware efficient, inter-neuronal transfer functions (or kernels) which carry information in the form of postsynaptic membrane potentials, are the most important (and erstwhile missing) element in neuromorphic implementations of Spiking Neural Networks (SNN). In the absence of such abstract kernels, spiking neuromorphic systems must realize very large numbers of synapses and their associated connectivity. The resultant hardware and bandwidth limitations create difficult tradeoffs which diminish the usefulness of such systems.
In this thesis a novel model of spiking neurons is proposed. The proposed Synapto-dendritic Kernel Adapting Neuron (SKAN) uses the adaptation of their synapto-dendritic kernels in conjunction with an adaptive threshold to perform unsupervised learning and inference on spatio-temporal spike patterns. The hardware and connectivity requirements of the neuron model are minimized through the use of simple accumulator-based kernels as well as through the use of timing information to perform a winner take all operation between the neurons. The learning and inference operations of SKAN are characterized and shown to be robust across a range of noise environments.
Next, the SKAN model is augmented with a simplified hardware-efficient model of Spike Timing Dependent Plasticity (STDP). In biology STDP is the mechanism which allows neurons to learn spatio-temporal spike patterns. However when the proposed SKAN model is augmented with a simplified STDP rule, where the synaptic kernel is used as a binary flag that enable synaptic potentiation, the result is a synaptic encoding of afferent Signal to Noise Ratio (SNR). In this combined model the neuron not only learns the target spatio-temporal spike patterns but also weighs each channel independently according to its signal to noise ratio. Additionally a novel approach is presented to achieving homeostatic plasticity in digital hardware which reduces hardware cost by eliminating the need for multipliers.
Finally the behavior and potential utility of this combined model is investigated in a range of noise conditions and the digital hardware resource utilization of SKAN and SKAN + STDP is detailed using Field Programmable Gate Arrays (FPGA)
Hardware-Amenable Structural Learning for Spike-based Pattern Classification using a Simple Model of Active Dendrites
This paper presents a spike-based model which employs neurons with
functionally distinct dendritic compartments for classifying high dimensional
binary patterns. The synaptic inputs arriving on each dendritic subunit are
nonlinearly processed before being linearly integrated at the soma, giving the
neuron a capacity to perform a large number of input-output mappings. The model
utilizes sparse synaptic connectivity; where each synapse takes a binary value.
The optimal connection pattern of a neuron is learned by using a simple
hardware-friendly, margin enhancing learning algorithm inspired by the
mechanism of structural plasticity in biological neurons. The learning
algorithm groups correlated synaptic inputs on the same dendritic branch. Since
the learning results in modified connection patterns, it can be incorporated
into current event-based neuromorphic systems with little overhead. This work
also presents a branch-specific spike-based version of this structural
plasticity rule. The proposed model is evaluated on benchmark binary
classification problems and its performance is compared against that achieved
using Support Vector Machine (SVM) and Extreme Learning Machine (ELM)
techniques. Our proposed method attains comparable performance while utilizing
10 to 50% less computational resources than the other reported techniques.Comment: Accepted for publication in Neural Computatio
Neuro-memristive Circuits for Edge Computing: A review
The volume, veracity, variability, and velocity of data produced from the
ever-increasing network of sensors connected to Internet pose challenges for
power management, scalability, and sustainability of cloud computing
infrastructure. Increasing the data processing capability of edge computing
devices at lower power requirements can reduce several overheads for cloud
computing solutions. This paper provides the review of neuromorphic
CMOS-memristive architectures that can be integrated into edge computing
devices. We discuss why the neuromorphic architectures are useful for edge
devices and show the advantages, drawbacks and open problems in the field of
neuro-memristive circuits for edge computing
An investigation into adaptive power reduction techniques for neural hardware
In light of the growing applicability of Artificial Neural Network (ANN) in the signal processing field [1] and the present thrust of the semiconductor industry towards lowpower SOCs for mobile devices [2], the power consumption of ANN hardware has become a very important implementation issue. Adaptability is a powerful and useful feature of neural networks. All current approaches for low-power ANN hardware techniques are ‘non-adaptive’ with respect to the power consumption of the network (i.e. power-reduction is not an objective of the adaptation/learning process). In the research work presented in this thesis, investigations on possible adaptive power reduction techniques have been carried out, which attempt to exploit the adaptability of neural networks in order to reduce the power consumption. Three separate approaches for such adaptive power reduction are proposed: adaptation of size, adaptation of network weights and adaptation of calculation precision. Initial case studies exhibit promising results with significantpower reduction
GPUs outperform current HPC and neuromorphic solutions in terms of speed and energy when simulating a highly-connected cortical model
While neuromorphic systems may be the ultimate platform for deploying spiking neural networks (SNNs), their distributed nature and optimisation for specific types of models makes them unwieldy tools for developing them. Instead, SNN models tend to be developed and simulated on computers or clusters of computers with standard von Neumann CPU architectures. Over the last decade, as well as becoming a common fixture in many workstations, NVIDIA GPU accelerators have entered the High Performance Computing field and are now used in 50% of the Top 10 super computing sites worldwide. In this paper we use our GeNN code generator to re-implement two neo-cortex-inspired, circuit-scale, point neuron network models on GPU hardware. We verify the correctness of our GPU simulations against prior results obtained with NEST running on traditional HPC hardware and compare the performance with respect to speed and energy consumption against published data from CPU-based HPC and neuromorphic hardware. A full-scale model of a cortical column can be simulated at speeds approaching 0.5× real-time using a single NVIDIA Tesla V100 accelerator – faster than is currently possible using a CPU based cluster or the SpiNNaker neuromorphic system. In addition, we find that, across a range of GPU systems, the energy to solution as well as the energy per synaptic event of the microcircuit simulation is as much as 14× lower than either on SpiNNaker or in CPU-based simulations. Besides performance in terms of speed and energy consumption of the simulation, efficient initialisation of models is also a crucial concern, particularly in a research context where repeated runs and parameter-space exploration are required. Therefore, we also introduce in this paper some of the novel parallel initialisation methods implemented in the latest version of GeNN and demonstrate how they can enable further speed and energy advantages
Real-time FGPA implementation of a neuromorphic pitch detection system
This thesis explores the real-time implementation of a biologically inspired pitch
detection system in digital electronics. Pitch detection is well understood and has been
shown to occur in the initial stages of the auditory brainstem. By building such a
system in digital hardware we can prove the feasibility of implementing neuromorphic
systems using digital technology.
This research not only aims to prove that such an implementation is possible but to
investigate ways of achieving efficient and effective designs. We aim to achieve this
complexity reduction while maintaining the fine granularity of the signal processing
inherent in neural systems. By producing an efficient design we present the possibility
of implementing the system within the available resources, thus producing a
demonstrable system. This thesis presents a review of computational models of all the
components within the pitch detection system. The review also identifies key issues
relating to the efficient implementation and development of the pitch detection
system. Four investigations are presented to address these issues for optimal
neuromorphic designs of neuromorphic systems.
The first investigation aims to produce the first-ever digital hardware implementation
of the inner hair cell. The second investigation develops simplified models of the
auditory nerve and the coincidence cell. The third investigation aims to reduce the
most complex stage of the system, the stellate chopper cell array. Finally, we
investigate implementing a large portion of the pitch detection system in hardware.
The results contained in this thesis enable us to understand the feasibility of
implementing such systems in real-time digital hardware. This knowledge may help
researchers to make design decisions within the field of digital neuromorphic systems
A Complete Arithmetic Calculator Constructed from Spiking Neural P Systems and its Application to Information Fusion
© World Scientific Publishing Company Several variants of spiking neural P systems (SNPS) have been presented in the literature to perform arithmetic operations. However, each of these variants was designed only for one specific arithmetic operation. In this paper, a complete arithmetic calculator implemented by SNPS is proposed. An application of the proposed calculator to information fusion is also proposed. The information fusion is implemented by integrating the following three elements: (1) an addition and subtraction SNPS already reported in the literature; (2) a modified multiplication and division SNPS; (3) a novel storage SNPS, i.e. a method based on SNPS is introduced to calculate basic probability assignment of an event. This is the first attempt to apply arithmetic operation SNPS to fuse multiple information. The effectiveness of the presented general arithmetic SNPS calculator is verified by means of several examples
Communication sparsity in distributed spiking neural network simulations to improve scalability
In the last decade there has been a surge in the number of big science projects interested in achieving a comprehensive understanding of the functions of the brain, using Spiking Neuronal Network (SNN) simulations to aid discovery and experimentation. Such an approach increases the computational demands on SNN simulators: if natural scale brain-size simulations are to be realized, it is necessary to use parallel and distributed models of computing. Communication is recognized as the dominant part of distributed SNN simulations. As the number of computational nodes increases, the proportion of time the simulation spends in useful computing (computational efficiency) is reduced and therefore applies a limit to scalability. This work targets the three phases of communication to improve overall computational efficiency in distributed simulations: implicit synchronization, process handshake and data exchange. We introduce a connectivity-aware allocation of neurons to compute nodes by modeling the SNN as a hypergraph. Partitioning the hypergraph to reduce interprocess communication increases the sparsity of the communication graph. We propose dynamic sparse exchange as an improvement over simple point-to-point exchange on sparse communications. Results show a combined gain when using hypergraph-based allocation and dynamic sparse communication, increasing computational efficiency by up to 40.8 percentage points and reducing simulation time by up to 73%. The findings are applicable to other distributed complex system simulations in which communication is modeled as a graph network
Neural architecture for echo suppression during sound source localization based on spiking neural cell models
Zusammenfassung
Diese Arbeit untersucht die biologischen Ursachen des psycho-akustischen
Präzedenz Effektes, der Menschen in die Lage versetzt, akustische Echos während
der Lokalisation von Schallquellen zu unterdrücken. Sie enthält ein Modell zur
Echo-Unterdrückung während der Schallquellenlokalisation, welches in technischen
Systemen zur Mensch-Maschine Interaktion eingesetzt werden kann.
Die Grundlagen dieses Modells wurden aus eigenen elektrophysiologischen
Experimenten an der Mongolischen Wüstenrennmaus gewonnen. Die dabei erstmalig an
der Wüstenrennmaus erzielten Ergebnisse, zeigen ein besonderes Verhalten
spezifischer Zellen im Dorsalen Kern des Lateral Lemniscus, einer dedizierten
Region des auditorischen Hirnstammes. Die dort sichtbare Langzeithemmung scheint
die Grundlage für die Echounterdrückung in höheren auditorischen Zentren zu
sein. Das entwickelte Model war in der Lage dieses Verhalten nachzubilden, und
legt die Vermutung nahe, dass eine starke und zeitlich präzise Hyperpolarisation
der zugrundeliegende physiologische Mechanismus dieses Verhaltens ist.
Die entwickelte Neuronale Modellarchitektur modelliert das Innenohr und fünf
wesentliche Kerne des auditorischen Hirnstammes in ihrer Verbindungsstruktur und
internen Dynamik. Sie stellt einen neuen Typus neuronaler Modellierung dar, der
als Spike-Interaktionsmodell (SIM) bezeichnet wird. SIM nutzen die präzise
räumlich-zeitliche Interaktion einzelner Aktionspotentiale (Spikes) für die
Kodierung und Verarbeitung neuronaler Informationen. Die Basis dafür bilden
Integrate-and-Fire Neuronenmodelle sowie Hebb'sche Synapsen, welche um speziell
entwickelte dynamische Kernfunktionen erweitert wurden. Das Modell ist in der
Lage, Zeitdifferenzen von 10 mykrosekunden zu detektieren und basiert auf den
Prinzipien der zeitlichen und räumlichen Koinzidenz sowie der präzisen lokalen
Inhibition.
Es besteht ausschließlich aus Elementen einer eigens entwickelten Neuronalen
Basisbibliothek (NBL) die speziell für die Modellierung verschiedenster Spike-
Interaktionsmodelle entworfen wurde. Diese Bibliothek erweitert die kommerziell
verfügbare dynamische Simulationsumgebung von MATLAB/SIMULINK um verschiedene
Modelle von Neuronen und Synapsen, welche die intrinsischen dynamischen
Eigenschaften von Nervenzellen nachbilden. Die Nutzung dieser Bibliothek
versetzt sowohl den Ingenieur als auch den Biologen in die Lage, eigene,
biologisch plausible, Modelle der neuronalen Informationsverarbeitung ohne
detaillierte Programmierkenntnisse zu entwickeln. Die grafische Oberfläche
ermöglicht strukturelle sowie parametrische Modifikationen und ist in der Lage,
den Zeitverlauf mikroskopischer Zellpotentiale aber auch makroskopischer
Spikemuster während und nach der Simulation darzustellen.
Zwei grundlegende Elemente der Neuronalen Basisbibliothek wurden zur
Implementierung als spezielle analog-digitale Schaltungen vorbereitet.
Erste Silizium Implementierungen durch das Team des DFG Graduiertenkollegs GRK
164 konnten die Möglichkeit einer vollparallelen on line Verarbeitung von
Schallsignalen nachweisen. Durch Zuhilfenahme des im GRK entwickelten
automatisierten Layout Generators wird es möglich, spezielle Prozessoren zur
Anwendung biologischer Verarbeitungsprinzipien in technischen Systemen zu
entwickeln. Diese Prozessoren unterscheiden sich grundlegend von den klassischen
von Neumann Prozessoren indem sie räumlich und zeitlich verteilte Spikemuster,
anstatt sequentieller binärer Werte zur Informationsrepräsentation nutzen. Sie
erweitern das digitale Kodierungsprinzip durch die Dimensionen des Raumes (2
dimensionale Nachbarschaft) der Zeit (Frequenz, Phase und Amplitude) sowie der
zeitlichen Dynamik analoger Potentialverläufe.
Diese Dissertation besteht aus sieben Kapiteln, welche den verschiedenen
Bereichen der Computational Neuroscience gewidmet sind.
Kapitel 1 beschreibt die Motivation dieser Arbeit welche aus der Absicht rühren,
biologische Prinzipien der Schallverarbeitung zu erforschen und für technische
Systeme während der Interaktion mit dem Menschen nutzbar zu machen. Zusätzlich
werden fünf Gründe für die Nutzung von Spike-Interaktionsmodellen angeführt
sowie deren neuartiger Charakter beschrieben.
Kapitel 2 führt die biologischen Prinzipien der Schallquellenlokalisation und
den psychoakustischen Präzedenz Effekt ein. Aktuelle Hypothesen zur Entstehung
dieses Effektes werden anhand ausgewählter experimenteller Ergebnisse
verschiedener Forschungsgruppen diskutiert.
Kapitel 3 beschreibt die entwickelte Neuronale Basisbibliothek und führt die
einzelnen neuronalen Simulationselemente ein. Es erklärt die zugrundeliegenden
mathematischen Funktionen der dynamischen Komponenten und beschreibt deren
generelle Einsetzbarkeit zur dynamischen Simulation spikebasierter Neuronaler
Netzwerke.
Kapitel 4 enthält ein speziell entworfenes Modell des auditorischen Hirnstammes
beginnend mit den Filterkaskaden zur Simulation des Innenohres, sich fortsetzend
über mehr als 200 Zellen und 400 Synapsen in 5 auditorischen Kernen bis zum
Richtungssensor im Bereich des auditorischen Mittelhirns. Es stellt die
verwendeten Strukturen und Parameter vor und enthält grundlegende Hinweise zur
Nutzung der Simulationsumgebung.
Kapitel 5 besteht aus drei Abschnitten, wobei der erste Abschnitt die
Experimentalbedingungen und Ergebnisse der eigens durchgeführten Tierversuche
beschreibt. Der zweite Abschnitt stellt die Ergebnisse von 104 Modellversuchen
zur Simulationen psycho-akustischer Effekte dar, welche u.a. die Fähigkeit des
Modells zur Nachbildung des Präzedenz Effektes testen. Schließlich beschreibt
der letzte Abschnitt die Ergebnisse der 54 unter realen Umweltbedingungen
durchgeführten Experimente. Dabei kamen Signale zur Anwendung, welche in
normalen sowie besonders stark verhallten Räumen aufgezeichnet wurden.
Kapitel 6 vergleicht diese Ergebnisse mit anderen biologisch motivierten und
technischen Verfahren zur Echounterdrückung und Schallquellenlokalisation und
führt den aktuellen Status der Hardwareimplementierung ein.
Kapitel 7 enthält schließlich eine kurze Zusammenfassung und einen Ausblick auf
weitere Forschungsobjekte und geplante Aktivitäten.
Diese Arbeit möchte zur Entwicklung der Computational Neuroscience beitragen,
indem sie versucht, in einem speziellen Anwendungsfeld die Lücke zwischen
biologischen Erkenntnissen, rechentechnischen Modellen und Hardware Engineering
zu schließen. Sie empfiehlt ein neues räumlich-zeitliches Paradigma der
dynamischen Informationsverarbeitung zur Erschließung biologischer Prinzipien
der Informationsverarbeitung für technische Anwendungen.This thesis investigates the biological background of the psycho-acoustical precedence effect, enabling humans to suppress echoes during the localization of sound sources. It provides a technically feasible and biologically plausible model for sound source localization under echoic conditions, ready to be used by technical systems during man-machine interactions.
The model is based upon own electro-physiological experiments in the mongolian gerbil. The first time in gerbils obtained results reveal a special behavior of specific cells of the dorsal nucleus of the lateral lemniscus (DNLL) - a distinct region in the auditory brainstem. The explored persistent inhibition effect of these cells seems to account for the base of echo suppression at higher auditory centers. The developed model proved capable to duplicate this behavior and suggests, that a strong and timely precise hyperpolarization is the basic mechanism behind this cell behavior. The developed neural architecture models the inner ear as well as five major nuclei of the auditory brainstem in their connectivity and intrinsic dynamics. It represents a new type of neural modeling described as Spike Interaction Models (SIM). SIM use the precise spatio-temporal interaction of single spike events for coding and processing of neural information. Their basic elements are Integrate-and-Fire Neurons and Hebbian synapses, which have been extended by specially designed dynamic transfer functions. The model is capable to detect time differences as small as 10 mircrosecondes and employs the principles of coincidence detection and precise local inhibition for auditory processing. It consists exclusively of elements of a specifically designed Neural Base Library (NBL), which has been developed for multi purpose modeling of Spike Interaction Models. This library extends the commercially available dynamic simulation environment of MATLAB/SIMULINK by different models of neurons and synapses simulating the intrinsic dynamic properties of neural cells. The usage of this library enables engineers as well as biologists to design their own, biologically plausible models of neural information processing without the need for detailed programming skills. Its graphical interface provides access to structural as well as parametric changes and is capable to display the time course of microscopic cell parameters as well as macroscopic firing pattern during simulations and thereafter. Two basic elements of the Neural Base Library have been prepared for implementation by specialized mixed analog-digital circuitry. First silicon implementations were realized by the team of the DFG Graduiertenkolleg GRK 164 and proved the possibility of fully parallel on line processing of sounds. By using the automated layout processor under development in the Graduiertenkolleg, it will be possible to design specific processors in order to apply theprinciples of distributed biological information processing to technical systems. These processors differ from classical von Neumann processors by the use of spatio temporal spike pattern instead of sequential binary values. They will extend the digital coding principle by the dimensions of space (spatial neighborhood), time (frequency, phase and amplitude) as well as the dynamics of analog potentials and introduce a new type of information processing.
This thesis consists of seven chapters, dedicated to the different areas of computational neuroscience.
Chapter 1: provides the motivation of this study arising from the attempt to investigate the biological principles of sound processing and make them available to technical systems interacting with humans under real world conditions. Furthermore, five reasons to use spike interaction models are given and their novel characteristics are discussed.
Chapter 2: introduces the biological principles of sound source localization and the precedence effect. Current hypothesis on echo suppression and the underlying principles of the precedence effect are discussed by reference to a small selection of physiological and psycho-acoustical experiments.
Chapter 3: describes the developed neural base library and introduces each of the designed neural simulation elements. It also explains the developed mathematical functions of the dynamic compartments and describes their general usage for dynamic simulation of spiking neural networks.
Chapter 4: introduces the developed specific model of the auditory brainstem, starting from the filtering cascade in the inner ear via more than 200 cells and 400 synapses in five auditory regions up to the directional sensor at the level of the auditory midbrain. It displays the employed parameter sets and contains basic hints for the set up and configuration of the simulation environment.
Chapter 5: consists of three sections, whereas the first one describes the set up and results of the own electro-physiological experiments. The second describes the results of 104 model simulations, performed to test the
models ability to duplicate psycho-acoustical effects like the precedence effect. Finally, the last section of this chapter contains the results of 54 real world experiments using natural sound signals, recorded under normal as well as highly reverberating conditions.
Chapter 6: compares the achieved results to other biologically motivated and technical models for echo suppression and sound source localization and introduces the current status of silicon implementation.
Chapter 7: finally provides a short summary and an outlook toward future research subjects and areas of investigation.
This thesis aims to contribute to the field of computational neuroscience by bridging the gap between biological
investigation, computational modeling and silicon engineering in a specific field of application. It suggests a new spatio-temporal paradigm of information processing in order to access the capabilities of biological systems for technical applications
- …