3 research outputs found
Evolutionary Dynamic Optimization Laboratory: A MATLAB Optimization Platform for Education and Experimentation in Dynamic Environments
Many real-world optimization problems possess dynamic characteristics.
Evolutionary dynamic optimization algorithms (EDOAs) aim to tackle the
challenges associated with dynamic optimization problems. Looking at the
existing works, the results reported for a given EDOA can sometimes be
considerably different. This issue occurs because the source codes of many
EDOAs, which are usually very complex algorithms, have not been made publicly
available. Indeed, the complexity of components and mechanisms used in many
EDOAs makes their re-implementation error-prone. In this paper, to assist
researchers in performing experiments and comparing their algorithms against
several EDOAs, we develop an open-source MATLAB platform for EDOAs, called
Evolutionary Dynamic Optimization LABoratory (EDOLAB). This platform also
contains an education module that can be used for educational purposes. In the
education module, the user can observe a) a 2-dimensional problem space and how
its morphology changes after each environmental change, b) the behaviors of
individuals over time, and c) how the EDOA reacts to environmental changes and
tries to track the moving optimum. In addition to being useful for research and
education purposes, EDOLAB can also be used by practitioners to solve their
real-world problems. The current version of EDOLAB includes 25 EDOAs and three
fully-parametric benchmark generators. The MATLAB source code for EDOLAB is
publicly available and can be accessed from
[https://github.com/EDOLAB-platform/EDOLAB-MATLAB].Comment: This work was submitted to ACM Transactions on Mathematical Software
on December 7, 202
An improved immune algorithm with parallel mutation and its application
The objective of this paper is to design a fast and efficient immune algorithm for solving various optimization problems. The immune algorithm (IA), which simulates the principle of the biological immune system, is one of the nature-inspired algorithms and its many advantages have been revealed. Although IA has shown its superiority over the traditional algorithms in many fields, it still suffers from the drawbacks of slow convergence and local minima trapping problems due to its inherent stochastic search property. Many efforts have been done to improve the search performance of immune algorithms, such as adaptive parameter setting and population diversity maintenance. In this paper, an improved immune algorithm (IIA) which utilizes a parallel mutation mechanism (PM) is proposed to solve the Lennard-Jones potential problem (LJPP). In IIA, three distinct mutation operators involving cauchy mutation (CM), gaussian mutation (GM) and lateral mutation (LM) are conditionally selected to be implemented. It is expected that IIA can effectively balance the exploration and exploitation of the search and thus speed up the convergence. To illustrate its validity, IIA is tested on a two-dimension function and some benchmark functions. Then IIA is applied to solve the LJPP to exhibit its applicability to the real-world problems. Experimental results demonstrate the effectiveness of IIA in terms of the convergence speed and the solution quality
Recommended from our members
HEDCOS: High Efficiency Dynamic Combinatorial Optimization System using Ant Colony Optimization algorithm
This thesis was submitted for the award of Doctor of Philosophy and was awarded by Brunel University LondonDynamic combinatorial optimization is gaining popularity among industrial practitioners due to the ever-increasing scale of their optimization problems and efforts to solve them to remain competitive. Larger optimization problems are not only more computationally intense to optimize but also have more uncertainty within problem inputs. If some aspects of the problem are subject to dynamic change, it becomes a Dynamic Optimization Problem (DOP).
In this thesis, a High Efficiency Dynamic Combinatorial Optimization System is built to solve challenging DOPs with high-quality solutions. The system is created using Ant Colony Optimization (ACO) baseline algorithm with three novel developments.
First, introduced an extension method for ACO algorithm called Dynamic Impact. Dynamic Impact is designed to improve convergence and solution quality by solving challenging optimization problems with a non-linear relationship between resource consumption and fitness. This proposed method is tested against the real-world Microchip Manufacturing Plant Production Floor Optimization (MMPPFO) problem and the theoretical benchmark Multidimensional Knapsack Problem (MKP).
Second, a non-stochastic dataset generation method was introduced to solve the dynamic optimization research replicability problem. This method uses a static benchmark dataset as a starting point and source of entropy to generate a sequence of dynamic states. Then using this method, 1405 Dynamic Multidimensional Knapsack Problem (DMKP) benchmark datasets were generated and published using famous static MKP benchmark instances as the initial state.
Third, introduced a nature-inspired discrete dynamic optimization strategy for ACO by modelling real-world ants’ symbiotic relationship with aphids. ACO with Aphids strategy is designed to solve discrete domain DOPs with event-triggered discrete dynamism. The strategy improved inter-state convergence by allowing better solution recovery after dynamic environment changes. Aphids mediate the information from previous dynamic optimization states to maximize initial results performance and minimize the impact on convergence speed. This strategy is tested for DMKP and against identical ACO implementations using Full-Restart and Pheromone-Sharing strategies, with all other variables isolated.
Overall, Dynamic Impact and ACO with Aphids developments are compounding. Using Dynamic Impact on single objective optimization of MMPPFO, the fitness value was improved by 33.2% over the ACO algorithm without Dynamic Impact. MKP benchmark instances of low complexity have been solved to a 100% success rate even when a high degree of solution sparseness is observed, and large complexity instances have shown the average gap improved by 4.26 times. ACO with Aphids has also demonstrated superior performance over the Pheromone-Sharing strategy in every test on average gap reduced by 29.2% for a total compounded dynamic optimization performance improvement of 6.02 times. Also, ACO with Aphids has outperformed the Full-Restart strategy for large datasets groups, and the overall average gap is reduced by 52.5% for a total compounded dynamic optimization performance improvement of 8.99 times