1,252 research outputs found
Comparison between Oja's and BCM neural networks models in finding useful projections in high-dimensional spaces
This thesis presents the concept of a neural network starting from its corresponding biological model, paying particular attention to the learning algorithms proposed by Oja and Bienenstock Cooper & Munro. A brief introduction to Data Analysis is then performed, with particular reference to the Principal Components Analysis and Singular Value Decomposition.
The two previously introduced algorithms are then dealt with more thoroughly, going to study in particular their connections with data analysis. Finally, it is proposed to use the Singular Value Decomposition as a method for obtaining stationary points in the BCM algorithm, in the case of linearly dependent inputs
Recommended from our members
Orbital Stability Analysis for Perturbed Nonlinear Systems and Natural Entrainment via Adaptive Andronov-Hopf Oscillator
New acceleration technique for the backpropagation algorithm
Artificial neural networks have been studied for many years in the hope of achieving human like performance in the area of pattern recognition, speech synthesis and higher level of cognitive process. In the connectionist model there are several interconnected processing elements called the neurons that have limited processing capability. Even though the rate of information transmitted between these elements is limited, the complex interconnection and the cooperative interaction between these elements results in a vastly increased computing power; The neural network models are specified by an organized network topology of interconnected neurons. These networks have to be trained in order them to be used for a specific purpose. Backpropagation is one of the popular methods of training the neural networks. There has been a lot of improvement over the speed of convergence of standard backpropagation algorithm in the recent past. Herein we have presented a new technique for accelerating the existing backpropagation without modifying it. We have used the fourth order interpolation method for the dominant eigen values, by using these we change the slope of the activation function. And by doing so we increase the speed of convergence of the backpropagation algorithm; Our experiments have shown significant improvement in the convergence time for problems widely used in benchmarKing Three to ten fold decrease in convergence time is achieved. Convergence time decreases as the complexity of the problem increases. The technique adjusts the energy state of the system so as to escape from local minima
In silico case studies of compliant robots: AMARSI deliverable 3.3
In the deliverable 3.2 we presented how the morphological computing ap-
proach can significantly facilitate the control strategy in several scenarios,
e.g. quadruped locomotion, bipedal locomotion and reaching. In particular,
the Kitty experimental platform is an example of the use of morphological
computation to allow quadruped locomotion. In this deliverable we continue
with the simulation studies on the application of the different morphological
computation strategies to control a robotic system
Multiplicative versus additive noise in multi-state neural networks
The effects of a variable amount of random dilution of the synaptic couplings
in Q-Ising multi-state neural networks with Hebbian learning are examined. A
fraction of the couplings is explicitly allowed to be anti-Hebbian. Random
dilution represents the dying or pruning of synapses and, hence, a static
disruption of the learning process which can be considered as a form of
multiplicative noise in the learning rule. Both parallel and sequential
updating of the neurons can be treated. Symmetric dilution in the statics of
the network is studied using the mean-field theory approach of statistical
mechanics. General dilution, including asymmetric pruning of the couplings, is
examined using the generating functional (path integral) approach of disordered
systems. It is shown that random dilution acts as additive gaussian noise in
the Hebbian learning rule with a mean zero and a variance depending on the
connectivity of the network and on the symmetry. Furthermore, a scaling factor
appears that essentially measures the average amount of anti-Hebbian couplings.Comment: 15 pages, 5 figures, to appear in the proceedings of the Conference
on Noise in Complex Systems and Stochastic Dynamics II (SPIE International
NASA JSC neural network survey results
A survey of Artificial Neural Systems in support of NASA's (Johnson Space Center) Automatic Perception for Mission Planning and Flight Control Research Program was conducted. Several of the world's leading researchers contributed papers containing their most recent results on artificial neural systems. These papers were broken into categories and descriptive accounts of the results make up a large part of this report. Also included is material on sources of information on artificial neural systems such as books, technical reports, software tools, etc
Non-Euclidean principal component analysis by Hebbian learning
Principal component analysis based on Hebbian learning is originally designed for data processing inEuclidean spaces. We present in this contribution an extension of Oja's Hebbian learning approach fornon-Euclidean spaces. We show that for Banach spaces the Hebbian learning can be carried out using theunderlying semi-inner product. Prominent examples for such Banach spaces are the lp-spaces for p≠2.For kernels spaces, as applied in support vector machines or kernelized vector quantization, thisapproach can be formulated as an online learning scheme based on the differentiable kernel. Hence,principal component analysis can be explicitly carried out in the respective data spaces but nowequipped with a non-Euclidean metric. In the article we provide the theoretical framework and giveillustrative examples
- …