86 research outputs found

    A Brief Survey on Intelligent Swarm-Based Algorithms for Solving Optimization Problems

    Get PDF
    This chapter presents an overview of optimization techniques followed by a brief survey on several swarm-based natural inspired algorithms which were introduced in the last decade. These techniques were inspired by the natural processes of plants, foraging behaviors of insects and social behaviors of animals. These swam intelligent methods have been tested on various standard benchmark problems and are capable in solving a wide range of optimization issues including stochastic, robust and dynamic problems

    Batch Reinforcement Learning on the Industrial Benchmark: First Experiences

    Full text link
    The Particle Swarm Optimization Policy (PSO-P) has been recently introduced and proven to produce remarkable results on interacting with academic reinforcement learning benchmarks in an off-policy, batch-based setting. To further investigate the properties and feasibility on real-world applications, this paper investigates PSO-P on the so-called Industrial Benchmark (IB), a novel reinforcement learning (RL) benchmark that aims at being realistic by including a variety of aspects found in industrial applications, like continuous state and action spaces, a high dimensional, partially observable state space, delayed effects, and complex stochasticity. The experimental results of PSO-P on IB are compared to results of closed-form control policies derived from the model-based Recurrent Control Neural Network (RCNN) and the model-free Neural Fitted Q-Iteration (NFQ). Experiments show that PSO-P is not only of interest for academic benchmarks, but also for real-world industrial applications, since it also yielded the best performing policy in our IB setting. Compared to other well established RL techniques, PSO-P produced outstanding results in performance and robustness, requiring only a relatively low amount of effort in finding adequate parameters or making complex design decisions

    Generating Interpretable Fuzzy Controllers using Particle Swarm Optimization and Genetic Programming

    Full text link
    Autonomously training interpretable control strategies, called policies, using pre-existing plant trajectory data is of great interest in industrial applications. Fuzzy controllers have been used in industry for decades as interpretable and efficient system controllers. In this study, we introduce a fuzzy genetic programming (GP) approach called fuzzy GP reinforcement learning (FGPRL) that can select the relevant state features, determine the size of the required fuzzy rule set, and automatically adjust all the controller parameters simultaneously. Each GP individual's fitness is computed using model-based batch reinforcement learning (RL), which first trains a model using available system samples and subsequently performs Monte Carlo rollouts to predict each policy candidate's performance. We compare FGPRL to an extended version of a related method called fuzzy particle swarm reinforcement learning (FPSRL), which uses swarm intelligence to tune the fuzzy policy parameters. Experiments using an industrial benchmark show that FGPRL is able to autonomously learn interpretable fuzzy policies with high control performance.Comment: Accepted at Genetic and Evolutionary Computation Conference 2018 (GECCO '18

    Escape Local Minima with Improved Particle Swarm Optimization Algorithm

    Get PDF
    Particle Swarm Optimization (PSO) is a powerful meta-heuristic technique which has been maneuvered to solve numerous complex optimization problems. However, due to its characteristics, there is a possibility to trap all particles in a local minimum in the solution space and then they cannot find the way out from the trap on their own. Therefore, we modify the traditional PSO algorithm by adding an extra step so that it helps PSO to find a better solution than the local minimum that they undesirably found. We perturb all the particles by adjusting parameter values in the traditional algorithm when there is no improvement of the objective value over the training iterations, assuming that particles have stuck in a local minimum. In this research, we mainly focus on adjusting the learning factors. However, the parameter values have to be used in an effective way to perturb the particles. The behavior of the proposed modification and its parameter adjustments are studied using a function which has a large number of local minima - Schwefel’s function. Results show that 2 out of 3 PSO attempts trap in local minimum and slight changes on learning factors do not help them to get out from the traps. However, perturbances made with large learning factors can find better solutions than the local minima that they stuck in and help to find the global minimum eventually

    A Brief Analysis of Gravitational Search Algorithm (GSA) Publication from 2009 to May 2013

    Get PDF
    Gravitational Search Algorithm was introduced in year 2009. Since its introduction, the academic community shows a great interest on this algorith. This can be seen by the high number of publications with a short span of time. This paper analyses the publication trend of Gravitational Search Algorithm since its introduction until May 2013. The objective of this paper is to give exposure to reader the publication trend in the area of Gravitational Search Algorithm
    • …
    corecore