912 research outputs found
Quantum Hopfield neural network
Quantum computing allows for the potential of significant advancements in
both the speed and the capacity of widely used machine learning techniques.
Here we employ quantum algorithms for the Hopfield network, which can be used
for pattern recognition, reconstruction, and optimization as a realization of a
content-addressable memory system. We show that an exponentially large network
can be stored in a polynomial number of quantum bits by encoding the network
into the amplitudes of quantum states. By introducing a classical technique for
operating the Hopfield network, we can leverage quantum algorithms to obtain a
quantum computational complexity that is logarithmic in the dimension of the
data. We also present an application of our method as a genetic sequence
recognizer.Comment: 13 pages, 3 figures, final versio
Neural Distributed Autoassociative Memories: A Survey
Introduction. Neural network models of autoassociative, distributed memory
allow storage and retrieval of many items (vectors) where the number of stored
items can exceed the vector dimension (the number of neurons in the network).
This opens the possibility of a sublinear time search (in the number of stored
items) for approximate nearest neighbors among vectors of high dimension. The
purpose of this paper is to review models of autoassociative, distributed
memory that can be naturally implemented by neural networks (mainly with local
learning rules and iterative dynamics based on information locally available to
neurons). Scope. The survey is focused mainly on the networks of Hopfield,
Willshaw and Potts, that have connections between pairs of neurons and operate
on sparse binary vectors. We discuss not only autoassociative memory, but also
the generalization properties of these networks. We also consider neural
networks with higher-order connections and networks with a bipartite graph
structure for non-binary data with linear constraints. Conclusions. In
conclusion we discuss the relations to similarity search, advantages and
drawbacks of these techniques, and topics for further research. An interesting
and still not completely resolved question is whether neural autoassociative
memories can search for approximate nearest neighbors faster than other index
structures for similarity search, in particular for the case of very high
dimensional vectors.Comment: 31 page
Optimisation in ‘Self-modelling’ Complex Adaptive Systems
When a dynamical system with multiple point attractors is released from an arbitrary initial condition it will relax into a configuration that locally resolves the constraints or opposing forces between interdependent state variables. However, when there are many conflicting interdependencies between variables, finding a configuration that globally optimises these constraints by this method is unlikely, or may take many attempts. Here we show that a simple distributed mechanism can incrementally alter a dynamical system such that it finds lower energy configurations, more reliably and more quickly. Specifically, when Hebbian learning is applied to the connections of a simple dynamical system undergoing repeated relaxation, the system will develop an associative memory that amplifies a subset of its own attractor states. This modifies the dynamics of the system such that its ability to find configurations that minimise total system energy, and globally resolve conflicts between interdependent variables, is enhanced. Moreover, we show that the system is not merely ‘recalling’ low energy states that have been previously visited but ‘predicting’ their location by generalising over local attractor states that have already been visited. This ‘self-modelling’ framework, i.e. a system that augments its behaviour with an associative memory of its own attractors, helps us better-understand the conditions under which a simple locally-mediated mechanism of self-organisation can promote significantly enhanced global resolution of conflicts between the components of a complex adaptive system. We illustrate this process in random and modular network constraint problems equivalent to graph colouring and distributed task allocation problems
APPLICATION OF NEURAL NETWORKS IN PREDICTIVE DATA MINING
Neural Networks represent a meaningfully different approach to using computers in the workplace. A neural network is used to learn patterns and relationships in data. The data may be the results of a market research effort, or the results of a production process given varying operational conditions. Regardless of the specifics involved, applying a neural network is a substantial departure from traditional approaches. In this paper we will look into how neural networks is used in data mining. The ultimate goal of data mining is prediction - and predictive data mining is the most common type of data mining and one that has the most direct business applications. Therefore, we will consider how this technique can be used to classify the performance status of a departmental store in monitoring their productsNeural networks, data mining, prediction
NASA JSC neural network survey results
A survey of Artificial Neural Systems in support of NASA's (Johnson Space Center) Automatic Perception for Mission Planning and Flight Control Research Program was conducted. Several of the world's leading researchers contributed papers containing their most recent results on artificial neural systems. These papers were broken into categories and descriptive accounts of the results make up a large part of this report. Also included is material on sources of information on artificial neural systems such as books, technical reports, software tools, etc
An Introduction to Quaternion-Valued Recurrent Projection Neural Networks
Hypercomplex-valued neural networks, including quaternion-valued neural
networks, can treat multi-dimensional data as a single entity. In this paper,
we introduce the quaternion-valued recurrent projection neural networks
(QRPNNs). Briefly, QRPNNs are obtained by combining the non-local projection
learning with the quaternion-valued recurrent correlation neural network
(QRCNNs). We show that QRPNNs overcome the cross-talk problem of QRCNNs. Thus,
they are appropriate to implement associative memories. Furthermore,
computational experiments reveal that QRPNNs exhibit greater storage capacity
and noise tolerance than their corresponding QRCNNs.Comment: Accepted to be Published in: Proceedings of the 8th Brazilian
Conference on Intelligent Systems (BRACIS 2019), October 15-18, 2019,
Salvador, BA, Brazi
Design of Oscillatory Neural Networks by Machine Learning
We demonstrate the utility of machine learning algorithms for the design of
Oscillatory Neural Networks (ONNs). After constructing a circuit model of the
oscillators in a machine-learning-enabled simulator and performing
Backpropagation through time (BPTT) for determining the coupling resistances
between the ring oscillators, we show the design of associative memories and
multi-layered ONN classifiers. The machine-learning-designed ONNs show superior
performance compared to other design methods (such as Hebbian learning) and
they also enable significant simplifications in the circuit topology. We
demonstrate the design of multi-layered ONNs that show superior performance
compared to single-layer ones. We argue Machine learning can unlock the true
computing potential of ONNs hardware
- …