1,230 research outputs found

    Metaheuristic design of feedforward neural networks: a review of two decades of research

    Get PDF
    Over the past two decades, the feedforward neural network (FNN) optimization has been a key interest among the researchers and practitioners of multiple disciplines. The FNN optimization is often viewed from the various perspectives: the optimization of weights, network architecture, activation nodes, learning parameters, learning environment, etc. Researchers adopted such different viewpoints mainly to improve the FNN's generalization ability. The gradient-descent algorithm such as backpropagation has been widely applied to optimize the FNNs. Its success is evident from the FNN's application to numerous real-world problems. However, due to the limitations of the gradient-based optimization methods, the metaheuristic algorithms including the evolutionary algorithms, swarm intelligence, etc., are still being widely explored by the researchers aiming to obtain generalized FNN for a given problem. This article attempts to summarize a broad spectrum of FNN optimization methodologies including conventional and metaheuristic approaches. This article also tries to connect various research directions emerged out of the FNN optimization practices, such as evolving neural network (NN), cooperative coevolution NN, complex-valued NN, deep learning, extreme learning machine, quantum NN, etc. Additionally, it provides interesting research challenges for future research to cope-up with the present information processing era

    The design and applications of the african buffalo algorithm for general optimization problems

    Get PDF
    Optimization, basically, is the economics of science. It is concerned with the need to maximize profit and minimize cost in terms of time and resources needed to execute a given project in any field of human endeavor. There have been several scientific investigations in the past several decades on discovering effective and efficient algorithms to providing solutions to the optimization needs of mankind leading to the development of deterministic algorithms that provide exact solutions to optimization problems. In the past five decades, however, the attention of scientists has shifted from the deterministic algorithms to the stochastic ones since the latter have proven to be more robust and efficient, even though they do not guarantee exact solutions. Some of the successfully designed stochastic algorithms include Simulated Annealing, Genetic Algorithm, Ant Colony Optimization, Particle Swarm Optimization, Bee Colony Optimization, Artificial Bee Colony Optimization, Firefly Optimization etc. A critical look at these ‘efficient’ stochastic algorithms reveals the need for improvements in the areas of effectiveness, the number of several parameters used, premature convergence, ability to search diverse landscapes and complex implementation strategies. The African Buffalo Optimization (ABO), which is inspired by the herd management, communication and successful grazing cultures of the African buffalos, is designed to attempt solutions to the observed shortcomings of the existing stochastic optimization algorithms. Through several experimental procedures, the ABO was used to successfully solve benchmark optimization problems in mono-modal and multimodal, constrained and unconstrained, separable and non-separable search landscapes with competitive outcomes. Moreover, the ABO algorithm was applied to solve over 100 out of the 118 benchmark symmetric and all the asymmetric travelling salesman’s problems available in TSPLIB95. Based on the successful experimentation with the novel algorithm, it is safe to conclude that the ABO is a worthy contribution to the scientific literature

    A survey of swarm intelligence for dynamic optimization: algorithms and applications

    Get PDF
    Swarm intelligence (SI) algorithms, including ant colony optimization, particle swarm optimization, bee-inspired algorithms, bacterial foraging optimization, firefly algorithms, fish swarm optimization and many more, have been proven to be good methods to address difficult optimization problems under stationary environments. Most SI algorithms have been developed to address stationary optimization problems and hence, they can converge on the (near-) optimum solution efficiently. However, many real-world problems have a dynamic environment that changes over time. For such dynamic optimization problems (DOPs), it is difficult for a conventional SI algorithm to track the changing optimum once the algorithm has converged on a solution. In the last two decades, there has been a growing interest of addressing DOPs using SI algorithms due to their adaptation capabilities. This paper presents a broad review on SI dynamic optimization (SIDO) focused on several classes of problems, such as discrete, continuous, constrained, multi-objective and classification problems, and real-world applications. In addition, this paper focuses on the enhancement strategies integrated in SI algorithms to address dynamic changes, the performance measurements and benchmark generators used in SIDO. Finally, some considerations about future directions in the subject are given

    Introductory Chapter: Swarm Intelligence and Particle Swarm Optimization

    Get PDF

    Stability Analysis of Artificial Bee Colony Optimization Algorithm

    Get PDF
    Theoretical analysis of swarm intelligence and evolutionary algorithms is relatively less explored area of research. Stability and convergence analysis of swarm intelligence and evolutionary algorithms can help the researchers to fine tune the parameter values. This paper presents the stability analysis of a famous Artificial Bee Colony (ABC) optimization algorithm using von Neumann stability criterion for two-level finite difference scheme. Parameter selection for the ABC algorithm is recommended based on the obtained stability conditions. The findings are also validated through numerical experiments on test problems

    BQIABC: A new Quantum-Inspired Artificial Bee Colony Algorithm for Binary Optimization Problems

    Get PDF
    Artificial bee colony (ABC) algorithm is a swarm intelligence optimization algorithm inspired by the intelligent behavior of honey bees when searching for food sources. The various versions of the ABC algorithm have been widely used to solve continuous and discrete optimization problems in different fields. In this paper a new binary version of the ABC algorithm inspired by quantum computing, called binary quantum-inspired artificial bee colony algorithm (BQIABC), is proposed. The BQIABC combines the main structure of ABC with the concepts and principles of quantum computing such as, quantum bit, quantum superposition state and rotation Q-gates strategy to make an algorithm with more exploration ability. The proposed algorithm due to its higher exploration ability can provide a robust tool to solve binary optimization problems. To evaluate the effectiveness of the proposed algorithm, several experiments are conducted on the 0/1 knapsack problem, Max-Ones and Royal-Road functions. The results produced by BQIABC are compared with those of ten state-of-the-art binary optimization algorithms. Comparisons show that BQIABC presents the better results than or similar to other algorithms. The proposed algorithm can be regarded as a promising algorithm to solve binary optimization problems
    corecore