5 research outputs found

    Effectiveness of firefly algorithm based neural network in time series forecasting

    Get PDF
    Global optimization techniques such as Particle Swarm Optimizers (PSO) and Genetic Algorithm (GA) are now widely used for training Artificial Neural Networks (NN), particularly in time series forecasting problems. Firefly algorithm (FA) is a relatively new addition to the family of population based optimization technique that has shown promising result in a number of problems. In this work, we evaluate the effectiveness of FA trained NN in time series forecasting. In the experiments, three well known time series were used to evaluate the performance. Results obtained were compared with results from both PSO and Resilient Propagation (RPROP) trained NNs. FA based NN performed very well in forecasting all the time series considered, outperforming the bench-marks in two out of the three problems.Keywords: Time series, Artificial Neural Network, Firefly Algorithm, Particle Swarm Optimization, Overfittin

    Parameter optimization of evolving spiking neural network with dynamic population particle swarm optimization

    Get PDF
    Evolving Spiking Neural Network (ESNN) is widely used in classification problem. However, ESNN like any other neural networks is incapable to find its own parameter optimum values, which are crucial for classification accuracy. Thus, in this study, ESNN is integrated with an improved Particle Swarm Optimization (PSO) known as Dynamic Population Particle Swarm Optimization (DPPSO) to optimize the ESNN parameters: the modulation factor (Mod), similarity factor (Sim) and threshold factor (C). To find the optimum ESNN parameter value, DPPSO uses a dynamic population that removes the lowest particle value in every pre-defined iteration. The integration of ESNN-DPPSO facilitates the ESNN parameter optimization searching during the training stage. The performance analysis is measured by classification accuracy and is compared with the existing method. Five datasets gained from University of California Irvine (UCI) Machine Learning Repository are used for this study. The experimental result presents better accuracy compared to the existing technique and thus improves the ESNN method in optimising its parameter values

    Parameter optimization of evolving spiking neural network with dynamic population particle swarm optimization

    Get PDF
    Evolving Spiking Neural Network (ESNN) is widely used in classification problem. However, ESNN like any other neural networks is incapable to find its own parameter optimum values, which are crucial for classification accuracy. Thus, in this study, ESNN is integrated with an improved Particle Swarm Optimization (PSO) known as Dynamic Population Particle Swarm Optimization (DPPSO) to optimize the ESNN parameters: the modulation factor (Mod), similarity factor (Sim) and threshold factor (C). To find the optimum ESNN parameter value, DPPSO uses a dynamic population that removes the lowest particle value in every pre-defined iteration. The integration of ESNN-DPPSO facilitates the ESNN parameter optimization searching during the training stage. The performance analysis is measured by classification accuracy and is compared with the existing method. Five datasets gained from University of California Irvine (UCI) Machine Learning Repository are used for this study. The experimental result presents better accuracy compared to the existing technique and thus improves the ESNN method in optimising its parameter values

    A survey of multi-population optimization algorithms for tracking the moving optimum in dynamic environments

    Get PDF
    The solution spaces of many real-world optimization problems change over time. Such problems are called dynamic optimization problems (DOPs), which pose unique challenges that necessitate adaptive strategies from optimization algorithms to maintain optimal performance and responsiveness to environmental changes. Tracking the moving optimum (TMO) is an important class of DOPs where the goal is to identify and deploy the best-found solution in each environments Multi-population dynamic optimization algorithms are particularly effective at solving TMOs due to their flexible structures and potential for adaptability. These algorithms are usually complex methods that are built by assembling multiple components, each of which is responsible for addressing a specific challenge or improving the tracking performance in response to changes. This survey provides an in-depth review of multi-population dynamic optimization algorithms, focusing on describing these algorithms as a set of multiple cooperating components, the synergy between these components, and their collective effectiveness and/or efficiency in addressing the challenges of TMOs. Additionally, this survey reviews benchmarking practices within this domain and outlines promising directions for future research

    Training neural networks with PSO in dynamic environments

    No full text
    corecore