6,014 research outputs found

    Genetic learning particle swarm optimization

    Get PDF
    Social learning in particle swarm optimization (PSO) helps collective efficiency, whereas individual reproduction in genetic algorithm (GA) facilitates global effectiveness. This observation recently leads to hybridizing PSO with GA for performance enhancement. However, existing work uses a mechanistic parallel superposition and research has shown that construction of superior exemplars in PSO is more effective. Hence, this paper first develops a new framework so as to organically hybridize PSO with another optimization technique for “learning.” This leads to a generalized “learning PSO” paradigm, the *L-PSO. The paradigm is composed of two cascading layers, the first for exemplar generation and the second for particle updates as per a normal PSO algorithm. Using genetic evolution to breed promising exemplars for PSO, a specific novel *L-PSO algorithm is proposed in the paper, termed genetic learning PSO (GL-PSO). In particular, genetic operators are used to generate exemplars from which particles learn and, in turn, historical search information of particles provides guidance to the evolution of the exemplars. By performing crossover, mutation, and selection on the historical information of particles, the constructed exemplars are not only well diversified, but also high qualified. Under such guidance, the global search ability and search efficiency of PSO are both enhanced. The proposed GL-PSO is tested on 42 benchmark functions widely adopted in the literature. Experimental results verify the effectiveness, efficiency, robustness, and scalability of the GL-PSO

    Optimizing Engagement Simulations Through the Advanced Framework for Simulation, Integration, and Modeling (AFSIM) Software

    Get PDF
    The ability to effectively model and simulate military missions holds the potential to save lives, money, and resources for the United States. The Advanced Framework for Simulation, Integration, and Modeling (AFSIM) software is a tool used to rapidly simulate and model new technologies and mission level scenarios. In this thesis, our objective is to integrate a closed loop optimization routine with AFSIM to identify an effective objective function to assess optimal inputs for engagement scenarios. Given the many factors which impact a mission level engagement, we developed a tool which interfaces with AFSIM to observe the effects from multiple inputs in an engagement scenario. Our tool operates under the assumption that simulation results have met an acceptable convergence threshold. The objective function evaluates the effectiveness and associated cost with a scenario using a genetic algorithm and a particle swarm optimization algorithm. From this, a statistical analysis was performed to assess risk from the distribution of effectiveness and cost at each point. The method allows an optimal set of inputs to be selected for a desired result from the selected engagement scenario.No embargoAcademic Major: Mechanical Engineerin

    From the social learning theory to a social learning algorithm for global optimization

    Get PDF
    Traditionally, the Evolutionary Computation (EC) paradigm is inspired by Darwinian evolution or the swarm intelligence of animals. Bandura's Social Learning Theory pointed out that the social learning behavior of humans indicates a high level of intelligence in nature. We found that such intelligence of human society can be implemented by numerical computing and be utilized in computational algorithms for solving optimization problems. In this paper, we design a novel and generic optimization approach that mimics the social learning process of humans. Emulating the observational learning and reinforcement behaviors, a virtual society deployed in the algorithm seeks the strongest behavioral patterns with the best outcome. This corresponds to searching for the best solution in solving optimization problems. Experimental studies in this paper showed the appealing search behavior of this human intelligence-inspired approach, which can reach the global optimum even in ill conditions. The effectiveness and high efficiency of the proposed algorithm has further been verified by comparing to some representative EC algorithms and variants on a set of benchmarks
    corecore