123 research outputs found
Recommended from our members
HEDCOS: High Efficiency Dynamic Combinatorial Optimization System using Ant Colony Optimization algorithm
This thesis was submitted for the award of Doctor of Philosophy and was awarded by Brunel University LondonDynamic combinatorial optimization is gaining popularity among industrial practitioners due to the ever-increasing scale of their optimization problems and efforts to solve them to remain competitive. Larger optimization problems are not only more computationally intense to optimize but also have more uncertainty within problem inputs. If some aspects of the problem are subject to dynamic change, it becomes a Dynamic Optimization Problem (DOP).
In this thesis, a High Efficiency Dynamic Combinatorial Optimization System is built to solve challenging DOPs with high-quality solutions. The system is created using Ant Colony Optimization (ACO) baseline algorithm with three novel developments.
First, introduced an extension method for ACO algorithm called Dynamic Impact. Dynamic Impact is designed to improve convergence and solution quality by solving challenging optimization problems with a non-linear relationship between resource consumption and fitness. This proposed method is tested against the real-world Microchip Manufacturing Plant Production Floor Optimization (MMPPFO) problem and the theoretical benchmark Multidimensional Knapsack Problem (MKP).
Second, a non-stochastic dataset generation method was introduced to solve the dynamic optimization research replicability problem. This method uses a static benchmark dataset as a starting point and source of entropy to generate a sequence of dynamic states. Then using this method, 1405 Dynamic Multidimensional Knapsack Problem (DMKP) benchmark datasets were generated and published using famous static MKP benchmark instances as the initial state.
Third, introduced a nature-inspired discrete dynamic optimization strategy for ACO by modelling real-world ants’ symbiotic relationship with aphids. ACO with Aphids strategy is designed to solve discrete domain DOPs with event-triggered discrete dynamism. The strategy improved inter-state convergence by allowing better solution recovery after dynamic environment changes. Aphids mediate the information from previous dynamic optimization states to maximize initial results performance and minimize the impact on convergence speed. This strategy is tested for DMKP and against identical ACO implementations using Full-Restart and Pheromone-Sharing strategies, with all other variables isolated.
Overall, Dynamic Impact and ACO with Aphids developments are compounding. Using Dynamic Impact on single objective optimization of MMPPFO, the fitness value was improved by 33.2% over the ACO algorithm without Dynamic Impact. MKP benchmark instances of low complexity have been solved to a 100% success rate even when a high degree of solution sparseness is observed, and large complexity instances have shown the average gap improved by 4.26 times. ACO with Aphids has also demonstrated superior performance over the Pheromone-Sharing strategy in every test on average gap reduced by 29.2% for a total compounded dynamic optimization performance improvement of 6.02 times. Also, ACO with Aphids has outperformed the Full-Restart strategy for large datasets groups, and the overall average gap is reduced by 52.5% for a total compounded dynamic optimization performance improvement of 8.99 times
A Brief Review on Mathematical Tools Applicable to Quantum Computing for Modelling and Optimization Problems in Engineering
Since its emergence, quantum computing has enabled a wide spectrum of new possibilities and advantages, including its efficiency in accelerating computational processes exponentially. This has directed much research towards completely novel ways of solving a wide variety of engineering problems, especially through describing quantum versions of many mathematical tools such as Fourier and Laplace transforms, differential equations, systems of linear equations, and optimization techniques, among others. Exploration and development in this direction will revolutionize the world of engineering. In this manuscript, we review the state of the art of these emerging techniques from the perspective of quantum computer development and performance optimization, with a focus on the most common mathematical tools that support engineering applications. This review focuses on the application of these mathematical tools to quantum computer development and performance improvement/optimization. It also identifies the challenges and limitations related to the exploitation of quantum computing and outlines the main opportunities for future contributions. This review aims at offering a valuable reference for researchers in fields of engineering that are likely to turn to quantum computing for solutions. Doi: 10.28991/ESJ-2023-07-01-020 Full Text: PD
Evolutionary Computation 2020
Intelligent optimization is based on the mechanism of computational intelligence to refine a suitable feature model, design an effective optimization algorithm, and then to obtain an optimal or satisfactory solution to a complex problem. Intelligent algorithms are key tools to ensure global optimization quality, fast optimization efficiency and robust optimization performance. Intelligent optimization algorithms have been studied by many researchers, leading to improvements in the performance of algorithms such as the evolutionary algorithm, whale optimization algorithm, differential evolution algorithm, and particle swarm optimization. Studies in this arena have also resulted in breakthroughs in solving complex problems including the green shop scheduling problem, the severe nonlinear problem in one-dimensional geodesic electromagnetic inversion, error and bug finding problem in software, the 0-1 backpack problem, traveler problem, and logistics distribution center siting problem. The editors are confident that this book can open a new avenue for further improvement and discoveries in the area of intelligent algorithms. The book is a valuable resource for researchers interested in understanding the principles and design of intelligent algorithms
A survey of swarm intelligence for dynamic optimization: algorithms and applications
Swarm intelligence (SI) algorithms, including ant colony optimization, particle swarm optimization, bee-inspired algorithms, bacterial foraging optimization, firefly algorithms, fish swarm optimization and many more, have been proven to be good methods to address difficult optimization problems under stationary environments. Most SI algorithms have been developed to address stationary optimization problems and hence, they can converge on the (near-) optimum solution efficiently. However, many real-world problems have a dynamic environment that changes over time. For such dynamic optimization problems (DOPs), it is difficult for a conventional SI algorithm to track the changing optimum once the algorithm has converged on a solution. In the last two decades, there has been a growing interest of addressing DOPs using SI algorithms due to their adaptation capabilities. This paper presents a broad review on SI dynamic optimization (SIDO) focused on several classes of problems, such as discrete, continuous, constrained, multi-objective and classification problems, and real-world applications. In addition, this paper focuses on the enhancement strategies integrated in SI algorithms to address dynamic changes, the performance measurements and benchmark generators used in SIDO. Finally, some considerations about future directions in the subject are given
A general framework of multi-population methods with clustering in undetectable dynamic environments
Copyright @ 2011 IEEETo solve dynamic optimization problems, multiple
population methods are used to enhance the population diversity for an algorithm with the aim of maintaining multiple populations in different sub-areas in the fitness landscape. Many experimental studies have shown that locating and tracking multiple relatively good optima rather than a single global optimum is an effective idea in dynamic environments. However, several challenges need to be addressed when multi-population methods are applied, e.g.,
how to create multiple populations, how to maintain them in different sub-areas, and how to deal with the situation where changes can not be detected or predicted. To address these issues, this paper investigates a hierarchical clustering method to locate and track multiple optima for dynamic optimization problems. To deal with undetectable dynamic environments, this
paper applies the random immigrants method without change detection based on a mechanism that can automatically reduce redundant individuals in the search space throughout the run. These methods are implemented into several research areas, including particle swarm optimization, genetic algorithm, and differential evolution. An experimental study is conducted based on the moving peaks benchmark to test the performance with several other algorithms from the literature. The experimental
results show the efficiency of the clustering method for locating and tracking multiple optima in comparison with other algorithms based on multi-population methods on the moving peaks
benchmark
Sine Cosine Algorithm for Optimization
This open access book serves as a compact source of information on sine cosine algorithm (SCA) and a foundation for developing and advancing SCA and its applications. SCA is an easy, user-friendly, and strong candidate in the field of metaheuristics algorithms. Despite being a relatively new metaheuristic algorithm, it has achieved widespread acceptance among researchers due to its easy implementation and robust optimization capabilities. Its effectiveness and advantages have been demonstrated in various applications ranging from machine learning, engineering design, and wireless sensor network to environmental modeling. The book provides a comprehensive account of the SCA, including details of the underlying ideas, the modified versions, various applications, and a working MATLAB code for the basic SCA
A Comprehensive Survey on Particle Swarm Optimization Algorithm and Its Applications
Particle swarm optimization (PSO) is a heuristic global optimization method, proposed originally by Kennedy and Eberhart in 1995. It is now one of the most commonly used optimization techniques. This survey presented a comprehensive investigation of PSO. On one hand, we provided advances with PSO, including its modifications (including quantum-behaved PSO, bare-bones PSO, chaotic PSO, and fuzzy PSO), population topology (as fully connected, von Neumann, ring, star, random, etc.), hybridization (with genetic algorithm, simulated annealing, Tabu search, artificial immune system, ant colony algorithm, artificial bee colony, differential evolution, harmonic search, and biogeography-based optimization), extensions (to multiobjective, constrained, discrete, and binary optimization), theoretical analysis (parameter selection and tuning, and convergence analysis), and parallel implementation (in multicore, multiprocessor, GPU, and cloud computing forms). On the other hand, we offered a survey on applications of PSO to the following eight fields: electrical and electronic engineering, automation control systems, communication theory, operations research, mechanical engineering, fuel and energy, medicine, chemistry, and biology. It is hoped that this survey would be beneficial for the researchers studying PSO algorithms
An improved MOEA/D algorithm for multi-objective multicast routing with network coding
Network coding enables higher network throughput, more balanced traffic, and securer data transmission. However, complicated mathematical operations incur when packets are combined at intermediate nodes, which, if not operated properly, lead to very high network resource consumption and unacceptable delay. Therefore, it is of vital importance to minimize various network resources and end-to-end delays while exploiting promising benefits of network coding.
Multicast has been used in increasingly more applications, such as video conferencing and remote education. In this paper the multicast routing problem with network coding is formulated as a multi-objective optimization problem (MOP), where the total coding cost, the total link cost and the end-to-end delay are minimized simultaneously. We adapt the multi-objective evolutionary algorithm based on decomposition (MOEA/D) for this MOP by hybridizing it with a population-based incremental learning technique which makes use of the global and historical information collected to provide additional guidance to the evolutionary search. Three new schemes are devised to facilitate the performance improvement, including a probability-based initialization scheme, a problem-specific population updating rule, and a hybridized reproduction operator. Experimental results clearly demonstrate that the proposed algorithm outperforms a number of state-of-the-art MOEAs regarding the solution quality and computational time
A parameter-free discrete particle swarm algorithm and its application to multi-objective pavement maintenance schemes
Regular maintenance is paramount for a healthy road network, the arteries of any economy. As the resources for
maintenance are limited, optimization is necessary. A number of conflicting objectives exist with many influencing
variables. Although many methods have been proposed, the related research is very active, due to difficulties
in adoption to the actual practice owing to reasons such high-dimensional problems even for small road
networks. Literature survey tells that particle swarms have not been exploited much, mainly due to unavailability
of many techniques in this domain for multi-objective discrete problems like this. In this work, a novel particle
swarm algorithm is proposed for a general, discrete, multi-objective problem. In contrast to the standard particle
swarm, the bare-bones technique has a clear advantage in that it is a parameter-free technique, hence the end
users need not be optimization experts. However, the existing barebones algorithm is available only for continuous
domains, sans any particle velocity terms. For discrete domains, the proposed method introduces a
parameter-free velocity term to the standard bare-bones algorithm. Based on the peak velocities observed by the
different dimensions of a particle, its new position is calculated. A number of benchmark test functions are also
solved. The results show that the proposed algorithm is highly competitive and able to obtain much better spread
of solutions compared to three other existing PSO and genetic algorithms. The method is benchmarked against a
number of other algorithms on an actual pavement maintenance problem. When compared against another
particle swarm algorithm, it not only shows better performance, but also significant reduction in run-time
compared to other POS algorithm. Hence, for large road network maintenance, the proposed method shows a lot of promise in terms of analysis time, while improving on the quality of solutions
- …