137 research outputs found
Connecting Spiking Neurons to a Spiking Memristor Network Changes the Memristor Dynamics
Memristors have been suggested as neuromorphic computing elements. Spike-time
dependent plasticity and the Hodgkin-Huxley model of the neuron have both been
modelled effectively by memristor theory. The d.c. response of the memristor is
a current spike. Based on these three facts we suggest that memristors are
well-placed to interface directly with neurons. In this paper we show that
connecting a spiking memristor network to spiking neuronal cells causes a
change in the memristor network dynamics by: removing the memristor spikes,
which we show is due to the effects of connection to aqueous medium; causing a
change in current decay rate consistent with a change in memristor state;
presenting more-linear dynamics; and increasing the memristor spiking
rate, as a consequence of interaction with the spiking neurons. This
demonstrates that neurons are capable of communicating directly with
memristors, without the need for computer translation.Comment: Conference paper, 4 page
Memristor Neural Network Design
Neural network, a powerful learning model, has archived amazing results. However, the current Von Neumann computing systemâbased implementations of neural networks are suffering from memory wall and communication bottleneck problems ascribing to the Complementary Metal Oxide Semiconductor (CMOS) technology scaling down and communication gap. Memristor, a two terminal nanosolid state nonvolatile resistive switching, can provide energyâefficient neuromorphic computing with its synaptic behavior. Crossbar architecture can be used to perform neural computations because of its high density and parallel computation. Thus, neural networks based on memristor crossbar will perform better in real world applications. In this chapter, the design of different neural network architectures based on memristor is introduced, including spiking neural networks, multilayer neural networks, convolution neural networks, and recurrent neural networks. And the brief introduction, the architecture, the computing circuits, and the training algorithm of each kind of neural networks are presented by instances. The potential applications and the prospects of memristorâbased neural network system are discussed
Recommended from our members
Versatile stochastic dot product circuits based on nonvolatile memories for high performance neurocomputing and neurooptimization.
The key operation in stochastic neural networks, which have become the state-of-the-art approach for solving problems in machine learning, information theory, and statistics, is a stochastic dot-product. While there have been many demonstrations of dot-product circuits and, separately, of stochastic neurons, the efficient hardware implementation combining both functionalities is still missing. Here we report compact, fast, energy-efficient, and scalable stochastic dot-product circuits based on either passively integrated metal-oxide memristors or embedded floating-gate memories. The circuit's high performance is due to mixed-signal implementation, while the efficient stochastic operation is achieved by utilizing circuit's noise, intrinsic and/or extrinsic to the memory cell array. The dynamic scaling of weights, enabled by analog memory devices, allows for efficient realization of different annealing approaches to improve functionality. The proposed approach is experimentally verified for two representative applications, namely by implementing neural network for solving a four-node graph-partitioning problem, and a Boltzmann machine with 10-input and 8-hidden neurons
- âŠ