28 research outputs found

    Memetic cooperative coevolution of Elman recurrent neural networks

    Get PDF
    Cooperative coevolution decomposes an optimi- sation problem into subcomponents and collectively solves them using evolutionary algorithms. Memetic algorithms provides enhancement to evolutionary algorithms with local search. Recently, the incorporation of local search into a memetic cooperative coevolution method has shown to be efficient for training feedforward networks on pattern classification problems. This paper applies the memetic cooperative coevolution method for training recurrent neural networks on grammatical inference problems. The results show that the proposed method achieves better performance in terms of optimisation time and robustness

    Problem Decomposition and Adaptation in Cooperative Neuro-Evolution

    No full text
    One way to train neural networks is to use evolutionary algorithms such as cooperative coevolution - a method that decomposes the network's learnable parameters into subsets, called subcomponents. Cooperative coevolution gains advantage over other methods by evolving particular subcomponents independently from the rest of the network. Its success depends strongly on how the problem decomposition is carried out. This thesis suggests new forms of problem decomposition, based on a novel and intuitive choice of modularity, and examines in detail at what stage and to what extent the different decomposition methods should be used. The new methods are evaluated by training feedforward networks to solve pattern classification tasks, and by training recurrent networks to solve grammatical inference problems. Efficient problem decomposition methods group interacting variables into the same subcomponents. We examine the methods from the literature and provide an analysis of the nature of the neural network optimization problem in terms of interacting variables. We then present a novel problem decomposition method that groups interacting variables and that can be generalized to neural networks with more than a single hidden layer. We then incorporate local search into cooperative neuro-evolution. We present a memetic cooperative coevolution method that takes into account the cost of employing local search across several sub-populations. The optimisation process changes during evolution in terms of diversity and interacting variables. To address this, we examine the adaptation of the problem decomposition method during the evolutionary process. The results in this thesis show that the proposed methods improve performance in terms of optimization time, scalability and robustness. As a further test, we apply the problem decomposition and adaptive cooperative coevolution methods for training recurrent neural networks on chaotic time series problems. The proposed methods show better performance in terms of accuracy and robustness

    Competitive two - island cooperative co - evolution for training feedforward neural networks for pattern classification problems

    Get PDF
    In the application of cooperative coevolution for neuro-evolution, problem decomposition methods rely on architectural properties of the neural network to divide it into subcomponents. During every stage of the evolutionary process, different problem decomposition methods yield unique characteristics that may be useful in an environment that enables solution sharing. In this paper, we implement a two-island competition environment in cooperative coevolution based neuro-evolution for feedforward neural networks for pattern classification problems. In particular the combinations of three problem decomposition methods that are based on the architectural properties that refers to neural level, network level and layer level decomposition. The experimental results show that the performance of the competition method is better than that of the standalone problem decomposition cooperative neuro-evolution methods

    Metaheuristic design of feedforward neural networks: a review of two decades of research

    Get PDF
    Over the past two decades, the feedforward neural network (FNN) optimization has been a key interest among the researchers and practitioners of multiple disciplines. The FNN optimization is often viewed from the various perspectives: the optimization of weights, network architecture, activation nodes, learning parameters, learning environment, etc. Researchers adopted such different viewpoints mainly to improve the FNN's generalization ability. The gradient-descent algorithm such as backpropagation has been widely applied to optimize the FNNs. Its success is evident from the FNN's application to numerous real-world problems. However, due to the limitations of the gradient-based optimization methods, the metaheuristic algorithms including the evolutionary algorithms, swarm intelligence, etc., are still being widely explored by the researchers aiming to obtain generalized FNN for a given problem. This article attempts to summarize a broad spectrum of FNN optimization methodologies including conventional and metaheuristic approaches. This article also tries to connect various research directions emerged out of the FNN optimization practices, such as evolving neural network (NN), cooperative coevolution NN, complex-valued NN, deep learning, extreme learning machine, quantum NN, etc. Additionally, it provides interesting research challenges for future research to cope-up with the present information processing era

    Language impairment and colour categories

    Get PDF
    Goldstein (1948) reported multiple cases of failure to categorise colours in patients that he termed amnesic or anomic aphasics. these patients have a particular difficulty in producing perceptual categories in the absence of other aphasic impairments. we hold that neuropsychological evidence supports the view that the task of colour categorisation is logically impossible without labels

    How culture might constrain color categories

    Get PDF
    If language is crucial to the development of shared colour categories, how might cultural constraints influence the development of divergent category sets? We propose that communities arrive at different sets of categories because the tendency to group by perceptual similarity interacts with environmental factors (differential access to dying and printing technologies), to make different systems optimal for communication in different situations.</p

    Turing Learning: Advances and Applications

    Get PDF
    Turing Learning is the family of algorithms where models and discriminators are generated in a competitive setting. This thesis concerns the coevolutionary framework of Turing Learning and investigates the advances for improved model accuracy and the applications in robotic systems. Advances proposed in this thesis are as follows: an interactive approach to enable the discriminator to genuinely influence the data sampling process; a hybrid formulation to combine the benefits of the interactive discriminator in improving model accuracy and the advantages of the passive discriminator for reducing training cost; an exclusiveness reward mechanism to promote candidates with the exclusive performance during the coevolutionary process. Applications presented in this thesis are as follows: an approach for a mobile robotic agent to automatically infer its sensor configuration; an approach for the robot agent to automatically calibrate its sensor reading; a novel approach to infer swarm behaviours from their effects on the environment. The interactive approach has been validated in the inference of sensor configuration and calibration model, leading to the self-modelling/self-discovery process of robotic agents. Results suggest an improved model accuracy with the interactive approach in both cases, compared with the passive approach. The hybrid formulation and the exclusiveness reward mechanism have been demonstrated in the inference of the calibration model. Results show that almost half of the training cost can be reduced without a decrease in model accuracy by applying the hybrid formulation. The novel reward mechanism can accelerate the convergence without a decrease in model accuracy. The indirect way of inferring swarm behaviours requires a small amount of training and reveals novel behavioural controllers for individual robots
    corecore