96,679 research outputs found
Metaheuristic Algorithms for Convolution Neural Network
A typical modern optimization technique is usually either heuristic or
metaheuristic. This technique has managed to solve some optimization problems
in the research area of science, engineering, and industry. However,
implementation strategy of metaheuristic for accuracy improvement on
convolution neural networks (CNN), a famous deep learning method, is still
rarely investigated. Deep learning relates to a type of machine learning
technique, where its aim is to move closer to the goal of artificial
intelligence of creating a machine that could successfully perform any
intellectual tasks that can be carried out by a human. In this paper, we
propose the implementation strategy of three popular metaheuristic approaches,
that is, simulated annealing, differential evolution, and harmony search, to
optimize CNN. The performances of these metaheuristic methods in optimizing CNN
on classifying MNIST and CIFAR dataset were evaluated and compared.
Furthermore, the proposed methods are also compared with the original CNN.
Although the proposed methods show an increase in the computation time, their
accuracy has also been improved (up to 7.14 percent).Comment: Article ID 1537325, 13 pages. Received 29 January 2016; Revised 15
April 2016; Accepted 10 May 2016. Academic Editor: Martin Hagan. in Hindawi
Publishing. Computational Intelligence and Neuroscience Volume 2016 (2016
Evolving Plasticity for Autonomous Learning under Changing Environmental Conditions
A fundamental aspect of learning in biological neural networks is the
plasticity property which allows them to modify their configurations during
their lifetime. Hebbian learning is a biologically plausible mechanism for
modeling the plasticity property in artificial neural networks (ANNs), based on
the local interactions of neurons. However, the emergence of a coherent global
learning behavior from local Hebbian plasticity rules is not very well
understood. The goal of this work is to discover interpretable local Hebbian
learning rules that can provide autonomous global learning. To achieve this, we
use a discrete representation to encode the learning rules in a finite search
space. These rules are then used to perform synaptic changes, based on the
local interactions of the neurons. We employ genetic algorithms to optimize
these rules to allow learning on two separate tasks (a foraging and a
prey-predator scenario) in online lifetime learning settings. The resulting
evolved rules converged into a set of well-defined interpretable types, that
are thoroughly discussed. Notably, the performance of these rules, while
adapting the ANNs during the learning tasks, is comparable to that of offline
learning methods such as hill climbing.Comment: Evolutionary Computation Journa
- …