113 research outputs found

    Global identification of electrical and mechanical parameters in PMSM drive based on dynamic self-learning PSO

    Get PDF
    A global parameter estimation method for a PMSM drive system is proposed, where the electrical parameters, mechanical parameters and voltage-source-inverter (VSI) nonlinearity are regarded as a whole and parameter estimation is formulated as a single parameter optimization model. A dynamic learning estimator is proposed for tracking the electrical parameters, mechanical parameters and VSI of PMSM drive by using dynamic self learning particle swarm optimization (DSLPSO). In DSLPSO, a novel movement modification equation with dynamic exemplar learning strategy is designed to ensure its diversity and achieve a reasonable tradeoff between the exploitation and exploration during the search process. Moreover, a nonlinear multi-scale based interactive learning operator is introduced for accelerating the convergence speed of the Pbest particles; meanwhile a dynamic opposition-based learning (OBL) strategy is designed to facilitate the gBest particle to explore a potentially better region. The proposed algorithm is applied to parameter estimation for a PMSM drive system. The results show that the proposed method has better performance in tracking the variation of electrical parameters, and estimating the immeasurable mechanical parameters and the VSI disturbance voltage simultaneously

    Population based optimization algorithms improvement using the predictive particles

    Get PDF
    A new efficient improvement, called Predictive Particle Modification (PPM), is proposed in this paper. This modification makes the particle look to the near area before moving toward the best solution of the group. This modification can be applied to any population algorithm. The basic philosophy of PPM is explained in detail. To evaluate the performance of PPM, it is applied to Particle Swarm Optimization (PSO) algorithm and Teaching Learning Based Optimization (TLBO) algorithm then tested using 23 standard benchmark functions. The effectiveness of these modifications are compared with the other unmodified population optimization algorithms based on the best solution, average solution, and convergence rate

    Feature selection using enhanced particle swarm optimisation for classification models.

    Get PDF
    In this research, we propose two Particle Swarm Optimisation (PSO) variants to undertake feature selection tasks. The aim is to overcome two major shortcomings of the original PSO model, i.e., premature convergence and weak exploitation around the near optimal solutions. The first proposed PSO variant incorporates four key operations, including a modified PSO operation with rectified personal and global best signals, spiral search based local exploitation, Gaussian distribution-based swarm leader enhancement, and mirroring and mutation operations for worst solution improvement. The second proposed PSO model enhances the first one through four new strategies, i.e., an adaptive exemplar breeding mechanism incorporating multiple optimal signals, nonlinear function oriented search coefficients, exponential and scattering schemes for swarm leader, and worst solution enhancement, respectively. In comparison with a set of 15 classical and advanced search methods, the proposed models illustrate statistical superiority for discriminative feature selection for a total of 13 data sets

    A dynamic neighborhood learning-based gravitational search algorithm

    Get PDF
    Balancing exploration and exploitation according to evolutionary states is crucial to meta-heuristic search (M-HS) algorithms. Owing to its simplicity in theory and effectiveness in global optimization, gravitational search algorithm (GSA) has attracted increasing attention in recent years. However, the tradeoff between exploration and exploitation in GSA is achieved mainly by adjusting the size of an archive, named Kbest, which stores those superior agents after fitness sorting in each iteration. Since the global property of Kbest remains unchanged in the whole evolutionary process, GSA emphasizes exploitation over exploration and suffers from rapid loss of diversity and premature convergence. To address these problems, in this paper, we propose a dynamic neighborhood learning (DNL) strategy to replace the Kbest model and thereby present a DNL-based GSA (DNLGSA). The method incorporates the local and global neighborhood topologies for enhancing the exploration and obtaining adaptive balance between exploration and exploitation. The local neighborhoods are dynamically formed based on evolutionary states. To delineate the evolutionary states, two convergence criteria named limit value and population diversity, are introduced. Moreover, a mutation operator is designed for escaping from the local optima on the basis of evolutionary states. The proposed algorithm was evaluated on 27 benchmark problems with different characteristic and various difficulties. The results reveal that DNLGSA exhibits competitive performances when compared with a variety of state-of-the-art M-HS algorithms. Moreover, the incorporation of local neighborhood topology reduces the numbers of calculations of gravitational force and thus alleviates the high computational cost of GSA

    GPU Implementation of DPSO-RE Algorithm for Parameters Identification of Surface PMSM Considering VSI Nonlinearity

    Get PDF
    In this paper, an accurate parameter estimation model of surface permanent magnet synchronous machines (SPMSMs) is established by taking into account voltage-source-inverter (VSI) nonlinearity. A fast dynamic particle swarm optimization (DPSO) algorithm combined with a receptor editing (RE) strategy is proposed to explore the optimal values of parameter estimations. This combination provides an accelerated implementation on graphics processing unit (GPU), and the proposed method is, therefore, referred to as G-DPSORE. In G-DPSO-RE, a dynamic labor division strategy is incorporated into the swarms according to the designed evolutionary factor during the evolution process. Two novel modifications of the movement equation are designed to update the velocity of particles. Moreover, a chaotic-logistic-based immune RE operator is developed to facilitate the global best individual (gBest particle) to explore a potentially better region. Furthermore, a GPU parallel acceleration technique is utilized to speed up parameter estimation procedure. It has been demonstrated that the proposed method is effective for simultaneous estimation of the PMSM parameters and the disturbance voltage (Vdead) due to VSI nonlinearity from experimental data for currents and rotor speed measured with inexpensive equipment. The influence of the VSI nonlinearity on the accuracy of parameter estimation is analyzed

    An Overview of Particle Swarm Optimization Variants

    Get PDF
    AbstractParticle swarm optimization (PSO) is a stochastic algorithm used for the optimization problems proposed by Kennedy [1] in 1995. It is a very good technique for the optimization problems. But still there is a drawback in the PSO is that it stuck in the local minima. To improve the performance of PSO, the researchers proposed the different variants of PSO. Some researchers try to improve it by improving initialization of the swarm. Some of them introduce the new parameters like constriction coefficient and inertia weight. Some researchers define the different method of inertia weight to improve the performance of PSO. Some researchers work on the global and local best particles by introducing the mutation operators in the PSO. In this paper, we will see the different variants of PSO with respect to initialization, inertia weight and mutation operators

    A Comprehensive Survey on Particle Swarm Optimization Algorithm and Its Applications

    Get PDF
    Particle swarm optimization (PSO) is a heuristic global optimization method, proposed originally by Kennedy and Eberhart in 1995. It is now one of the most commonly used optimization techniques. This survey presented a comprehensive investigation of PSO. On one hand, we provided advances with PSO, including its modifications (including quantum-behaved PSO, bare-bones PSO, chaotic PSO, and fuzzy PSO), population topology (as fully connected, von Neumann, ring, star, random, etc.), hybridization (with genetic algorithm, simulated annealing, Tabu search, artificial immune system, ant colony algorithm, artificial bee colony, differential evolution, harmonic search, and biogeography-based optimization), extensions (to multiobjective, constrained, discrete, and binary optimization), theoretical analysis (parameter selection and tuning, and convergence analysis), and parallel implementation (in multicore, multiprocessor, GPU, and cloud computing forms). On the other hand, we offered a survey on applications of PSO to the following eight fields: electrical and electronic engineering, automation control systems, communication theory, operations research, mechanical engineering, fuel and energy, medicine, chemistry, and biology. It is hoped that this survey would be beneficial for the researchers studying PSO algorithms
    • …
    corecore