114 research outputs found

    Feedback learning particle swarm optimization

    Get PDF
    This is the author’s version of a work that was accepted for publication in Applied Soft Computing. Changes resulting from the publishing process, such as peer review, editing, corrections, structural formatting, and other quality control mechanisms may not be reflected in this document. Changes may have been made to this work since it was submitted for publication. A definitive version was subsequently published and is available at the link below - Copyright @ Elsevier 2011In this paper, a feedback learning particle swarm optimization algorithm with quadratic inertia weight (FLPSO-QIW) is developed to solve optimization problems. The proposed FLPSO-QIW consists of four steps. Firstly, the inertia weight is calculated by a designed quadratic function instead of conventional linearly decreasing function. Secondly, acceleration coefficients are determined not only by the generation number but also by the search environment described by each particle’s history best fitness information. Thirdly, the feedback fitness information of each particle is used to automatically design the learning probabilities. Fourthly, an elite stochastic learning (ELS) method is used to refine the solution. The FLPSO-QIW has been comprehensively evaluated on 18 unimodal, multimodal and composite benchmark functions with or without rotation. Compared with various state-of-the-art PSO algorithms, the performance of FLPSO-QIW is promising and competitive. The effects of parameter adaptation, parameter sensitivity and proposed mechanism are discussed in detail.This research was partially supported by the National Natural Science Foundation of PR China (Grant No 60874113), the Research Fund for the Doctoral Program of Higher Education (Grant No 200802550007), the Key Creative Project of Shanghai Education Community (Grant No 09ZZ66), the Key Foundation Project of Shanghai(Grant No 09JC1400700), the International Science and Technology Cooperation Project of China under Grant 2009DFA32050, and the Alexander von Humboldt Foundation of Germany

    Adaptive particle swarm optimization

    Get PDF
    An adaptive particle swarm optimization (APSO) that features better search efficiency than classical particle swarm optimization (PSO) is presented. More importantly, it can perform a global search over the entire search space with faster convergence speed. The APSO consists of two main steps. First, by evaluating the population distribution and particle fitness, a real-time evolutionary state estimation procedure is performed to identify one of the following four defined evolutionary states, including exploration, exploitation, convergence, and jumping out in each generation. It enables the automatic control of inertia weight, acceleration coefficients, and other algorithmic parameters at run time to improve the search efficiency and convergence speed. Then, an elitist learning strategy is performed when the evolutionary state is classified as convergence state. The strategy will act on the globally best particle to jump out of the likely local optima. The APSO has comprehensively been evaluated on 12 unimodal and multimodal benchmark functions. The effects of parameter adaptation and elitist learning will be studied. Results show that APSO substantially enhances the performance of the PSO paradigm in terms of convergence speed, global optimality, solution accuracy, and algorithm reliability. As APSO introduces two new parameters to the PSO paradigm only, it does not introduce an additional design or implementation complexity

    Estimation of Stator Resistance and Rotor Flux Linkage in SPMSM Using CLPSO with Opposition-Based-Learning Strategy

    Get PDF
    Electromagnetic parameters are important for controller design and condition monitoring of permanent magnet synchronous machine (PMSM) system. In this paper, an improved comprehensive learning particle swarm optimization (CLPSO) with opposition-based-learning (OBL) strategy is proposed for estimating stator resistance and rotor flux linkage in surface-mounted PMSM; the proposed method is referred to as CLPSO-OBL. In the CLPSO-OBL framework, an opposition-learning strategy is used for best particles reinforcement learning to improve the dynamic performance and global convergence ability of the CLPSO. The proposed parameter optimization not only retains the advantages of diversity in the CLPSO but also has inherited global exploration capability of the OBL. Then, the proposed method is applied to estimate the stator resistance and rotor flux linkage of surface-mounted PMSM. The experimental results show that the CLPSO-OBL has better performance in estimating winding resistance and PM flux compared to the existing peer PSOs. Furthermore, the proposed parameter estimation model and optimization method are simple and with good accuracy, fast convergence, and easy digital implementation

    A self-learning particle swarm optimizer for global optimization problems

    Get PDF
    Copyright @ 2011 IEEE. All Rights Reserved. This article was made available through the Brunel Open Access Publishing Fund.Particle swarm optimization (PSO) has been shown as an effective tool for solving global optimization problems. So far, most PSO algorithms use a single learning pattern for all particles, which means that all particles in a swarm use the same strategy. This monotonic learning pattern may cause the lack of intelligence for a particular particle, which makes it unable to deal with different complex situations. This paper presents a novel algorithm, called self-learning particle swarm optimizer (SLPSO), for global optimization problems. In SLPSO, each particle has a set of four strategies to cope with different situations in the search space. The cooperation of the four strategies is implemented by an adaptive learning framework at the individual level, which can enable a particle to choose the optimal strategy according to its own local fitness landscape. The experimental study on a set of 45 test functions and two real-world problems show that SLPSO has a superior performance in comparison with several other peer algorithms.This work was supported by the Engineering and Physical Sciences Research Council of U.K. under Grants EP/E060722/1 and EP/E060722/2

    Orthogonal learning particle swarm optimization

    Get PDF
    Particle swarm optimization (PSO) relies on its learning strategy to guide its search direction. Traditionally, each particle utilizes its historical best experience and its neighborhood’s best experience through linear summation. Such a learning strategy is easy to use, but is inefficient when searching in complex problem spaces. Hence, designing learning strategies that can utilize previous search information (experience) more efficiently has become one of the most salient and active PSO research topics. In this paper, we proposes an orthogonal learning (OL) strategy for PSO to discover more useful information that lies in the above two experiences via orthogonal experimental design. We name this PSO as orthogonal learning particle swarm optimization (OLPSO). The OL strategy can guide particles to fly in better directions by constructing a much promising and efficient exemplar. The OL strategy can be applied to PSO with any topological structure. In this paper, it is applied to both global and local versions of PSO, yielding the OLPSO-G and OLPSOL algorithms, respectively. This new learning strategy and the new algorithms are tested on a set of 16 benchmark functions, and are compared with other PSO algorithms and some state of the art evolutionary algorithms. The experimental results illustrate the effectiveness and efficiency of the proposed learning strategy and algorithms. The comparisons show that OLPSO significantly improves the performance of PSO, offering faster global convergence, higher solution quality, and stronger robustness

    GPU Implementation of DPSO-RE Algorithm for Parameters Identification of Surface PMSM Considering VSI Nonlinearity

    Get PDF
    In this paper, an accurate parameter estimation model of surface permanent magnet synchronous machines (SPMSMs) is established by taking into account voltage-source-inverter (VSI) nonlinearity. A fast dynamic particle swarm optimization (DPSO) algorithm combined with a receptor editing (RE) strategy is proposed to explore the optimal values of parameter estimations. This combination provides an accelerated implementation on graphics processing unit (GPU), and the proposed method is, therefore, referred to as G-DPSORE. In G-DPSO-RE, a dynamic labor division strategy is incorporated into the swarms according to the designed evolutionary factor during the evolution process. Two novel modifications of the movement equation are designed to update the velocity of particles. Moreover, a chaotic-logistic-based immune RE operator is developed to facilitate the global best individual (gBest particle) to explore a potentially better region. Furthermore, a GPU parallel acceleration technique is utilized to speed up parameter estimation procedure. It has been demonstrated that the proposed method is effective for simultaneous estimation of the PMSM parameters and the disturbance voltage (Vdead) due to VSI nonlinearity from experimental data for currents and rotor speed measured with inexpensive equipment. The influence of the VSI nonlinearity on the accuracy of parameter estimation is analyzed

    Negatively Correlated Search

    Full text link
    Evolutionary Algorithms (EAs) have been shown to be powerful tools for complex optimization problems, which are ubiquitous in both communication and big data analytics. This paper presents a new EA, namely Negatively Correlated Search (NCS), which maintains multiple individual search processes in parallel and models the search behaviors of individual search processes as probability distributions. NCS explicitly promotes negatively correlated search behaviors by encouraging differences among the probability distributions (search behaviors). By this means, individual search processes share information and cooperate with each other to search diverse regions of a search space, which makes NCS a promising method for non-convex optimization. The cooperation scheme of NCS could also be regarded as a novel diversity preservation scheme that, different from other existing schemes, directly promotes diversity at the level of search behaviors rather than merely trying to maintain diversity among candidate solutions. Empirical studies showed that NCS is competitive to well-established search methods in the sense that NCS achieved the best overall performance on 20 multimodal (non-convex) continuous optimization problems. The advantages of NCS over state-of-the-art approaches are also demonstrated with a case study on the synthesis of unequally spaced linear antenna arrays

    Genetic learning particle swarm optimization

    Get PDF
    Social learning in particle swarm optimization (PSO) helps collective efficiency, whereas individual reproduction in genetic algorithm (GA) facilitates global effectiveness. This observation recently leads to hybridizing PSO with GA for performance enhancement. However, existing work uses a mechanistic parallel superposition and research has shown that construction of superior exemplars in PSO is more effective. Hence, this paper first develops a new framework so as to organically hybridize PSO with another optimization technique for “learning.” This leads to a generalized “learning PSO” paradigm, the *L-PSO. The paradigm is composed of two cascading layers, the first for exemplar generation and the second for particle updates as per a normal PSO algorithm. Using genetic evolution to breed promising exemplars for PSO, a specific novel *L-PSO algorithm is proposed in the paper, termed genetic learning PSO (GL-PSO). In particular, genetic operators are used to generate exemplars from which particles learn and, in turn, historical search information of particles provides guidance to the evolution of the exemplars. By performing crossover, mutation, and selection on the historical information of particles, the constructed exemplars are not only well diversified, but also high qualified. Under such guidance, the global search ability and search efficiency of PSO are both enhanced. The proposed GL-PSO is tested on 42 benchmark functions widely adopted in the literature. Experimental results verify the effectiveness, efficiency, robustness, and scalability of the GL-PSO

    Coevolutionary particle swarm optimization using AIS and its application in multiparameter estimation of PMSM

    Get PDF
    In this paper, a coevolutionary particle-swarm-optimization (PSO) algorithm associating with the artificial immune principle is proposed. In the proposed algorithm, the whole population is divided into two kinds of subpopulations consisting of one elite subpopulation and several normal subpopulations. The best individual of each normal subpopulation will be memorized into the elite subpopulation during the evolution process. A hybrid method, which creates new individuals by using three different operators, is presented to ensure the diversity of all the subpopulations. Furthermore, a simple adaptive wavelet learning operator is utilized for accelerating the convergence speed of the pbest particles. The improved immune-clonal-selection operator is employed for optimizing the elite subpopulation, while the migration scheme is employed for the information exchange between elite subpopulation and normal subpopulations. The performance of the proposed algorithm is verified by testing on a suite of standard benchmark functions, which shows faster convergence and global search ability. Its performance is further evaluated by its application to multiparameter estimation of permanent-magnet synchronous machines, which shows that its performance significantly outperforms existing PSOs. The proposed algorithm can estimate the machine dq-axis inductances, stator winding resistance, and rotor flux linkage simultaneously. © 2013 IEEE
    corecore