2,570 research outputs found
A Comparison of Nature Inspired Algorithms for Multi-threshold Image Segmentation
In the field of image analysis, segmentation is one of the most important
preprocessing steps. One way to achieve segmentation is by mean of threshold
selection, where each pixel that belongs to a determined class islabeled
according to the selected threshold, giving as a result pixel groups that share
visual characteristics in the image. Several methods have been proposed in
order to solve threshold selectionproblems; in this work, it is used the method
based on the mixture of Gaussian functions to approximate the 1D histogram of a
gray level image and whose parameters are calculated using three nature
inspired algorithms (Particle Swarm Optimization, Artificial Bee Colony
Optimization and Differential Evolution). Each Gaussian function approximates
thehistogram, representing a pixel class and therefore a threshold point.
Experimental results are shown, comparing in quantitative and qualitative
fashion as well as the main advantages and drawbacks of each algorithm, applied
to multi-threshold problem.Comment: 16 pages, this is a draft of the final version of the article sent to
the Journa
Differential evolution with an evolution path: a DEEP evolutionary algorithm
Utilizing cumulative correlation information already existing in an evolutionary process, this paper proposes a predictive approach to the reproduction mechanism of new individuals for differential evolution (DE) algorithms. DE uses a distributed model (DM) to generate new individuals, which is relatively explorative, whilst evolution strategy (ES) uses a centralized model (CM) to generate offspring, which through adaptation retains a convergence momentum. This paper adopts a key feature in the CM of a covariance matrix adaptation ES, the cumulatively learned evolution path (EP), to formulate a new evolutionary algorithm (EA) framework, termed DEEP, standing for DE with an EP. Without mechanistically combining two CM and DM based algorithms together, the DEEP framework offers advantages of both a DM and a CM and hence substantially enhances performance. Under this architecture, a self-adaptation mechanism can be built inherently in a DEEP algorithm, easing the task of predetermining algorithm control parameters. Two DEEP variants are developed and illustrated in the paper. Experiments on the CEC'13 test suites and two practical problems demonstrate that the DEEP algorithms offer promising results, compared with the original DEs and other relevant state-of-the-art EAs
Recommended from our members
Incremental evolution strategy for function optimization
This paper presents a novel evolutionary approach for function optimization Incremental Evolution Strategy (IES). Two strategies are proposed. One is to evolve the input variables incrementally. The whole evolution consists of several phases and one more variable is focused in each phase. The number of phases is equal to the number of variables in maximum. Each phase is composed of two stages: in the single-variable evolution (SVE) stage, evolution is taken on one independent variable in a series of cutting planes; in the multi-variable evolving (MVE) stage, the initial population is formed by integrating the populations obtained by the SVE and the MVE in the last phase. And the evolution is taken on the incremented variable set. The other strategy is a hybrid of particle swarm optimization (PSO) and evolution strategy (ES). PSO is applied to adjust the cutting planes/hyper-planes (in SVEs/MVEs) while (1+1)-ES is applied to searching optima in the cutting planes/hyper-planes. The results of experiments show that the performance of IES is generally better than that of three other evolutionary algorithms, improved normal GA, PSO and SADE_CERAF, in the sense that IES finds solutions closer to the true optima and with more optimal objective values
Proposal and Comparative Study of Evolutionary Algorithms for Optimum Design of a Gear System
This paper proposes a novel metaheuristic framework using a Differential Evolution (DE) algorithm with the Non-dominated Sorting Genetic Algorithm-II (NSGA-II). Both algorithms are combined employing a collaborative strategy with sequential execution, which is called DE-NSGA-II. The DE-NSGA-II takes advantage of the exploration abilities of the multi-objective evolutionary algorithms strengthened with the ability to search global mono-objective optimum of DE, that enhances the capability of finding those extreme solutions of Pareto Optimal Front (POF) difficult to achieve. Numerous experiments and performance comparisons between different evolutionary algorithms were performed on a referent problem for the mono-objective and multi-objective literature, which consists of the design of a double reduction gear train. A preliminary study of the problem, solved in an exhaustive way, discovers the low density of solutions in the vicinity of the optimal solution (mono-objective case) as well as in some areas of the POF of potential interest to a decision maker (multi-objective case). This characteristic of the problem would explain the considerable difficulties for its resolution when exact methods and/or metaheuristics are used, especially in the multi-objective case. However, the DE-NSGA-II framework exceeds these difficulties and obtains the whole POF which significantly improves the few previous multi-objective studies.Fil: MĂ©ndez Babey, Máximo. Universidad de Las Palmas de Gran Canaria; EspañaFil: Rossit, Daniel Alejandro. Universidad Nacional del Sur. Departamento de IngenierĂa; Argentina. Consejo Nacional de Investigaciones CientĂficas y TĂ©cnicas. Centro CientĂfico TecnolĂłgico Conicet - BahĂa Blanca. Instituto de Matemática BahĂa Blanca. Universidad Nacional del Sur. Departamento de Matemática. Instituto de Matemática BahĂa Blanca; ArgentinaFil: González, Begoña. Universidad de Las Palmas de Gran Canaria; EspañaFil: Frutos, Mariano. Universidad Nacional del Sur. Departamento de IngenierĂa; Argentina. Consejo Nacional de Investigaciones CientĂficas y TĂ©cnicas. Centro CientĂfico TecnolĂłgico Conicet - BahĂa Blanca. Instituto de Investigaciones EconĂłmicas y Sociales del Sur. Universidad Nacional del Sur. Departamento de EconomĂa. Instituto de Investigaciones EconĂłmicas y Sociales del Sur; Argentin
Adaptive intelligence applied to numerical optimisation
The article presents modification strategies theoretical comparison and experimental results achieved by adaptive heuristics applied to numerical optimisation of several non-constraint test functions. The aims of the study are to identify and compare how adaptive search heuristics behave within heterogeneous search space without retuning of the search parameters. The achieved results are summarised and analysed, which could be used for comparison to other methods and further investigation
Optimisation of Mobile Communication Networks - OMCO NET
The mini conference “Optimisation of Mobile Communication Networks” focuses on advanced methods for search and optimisation applied to wireless communication networks. It is sponsored by Research & Enterprise Fund Southampton Solent University.
The conference strives to widen knowledge on advanced search methods capable of optimisation of wireless communications networks. The aim is to provide a forum for exchange of recent knowledge, new ideas and trends in this progressive and challenging area. The conference will popularise new successful approaches on resolving hard tasks such as minimisation of transmit power, cooperative and optimal routing
Performance of Differential Evolution and Particle Swarm Methods on Some Relatively Harder Multi-modal Benchmark Functions
This paper aims at comparing the performance of the Differential Evolution (DE) and the Repulsive Particle Swarm (RPS) methods of global optimization. To this end, some relatively difficult test functions have been chosen. Among these test functions, some are new while others are well known in the literature. We use DE method with the exponential crossover scheme as well as with no crossover (only probabilistic replacement). Our findings suggest that DE (with the exponential crossover scheme) mostly fails to find the optimum in case of the functions under study. Of course, it succeeds in case of some functions (perm#2, zero-sum) for very small dimension, but begins to falter as soon as the dimension is increased. In case of DCS function, it works well up to dimension = 5. When we use no crossover (only probabilistic replacement) we obtain better results in case of several of the functions under study. In case of Perm#1, Perm#2, Zero-sum, Kowalik, Hougen and Power-sum functions, a remarkable advantage is there. Whether crossover or no crossover, DE falters when the optimand function has some element of randomness. This is indicated by the functions: Yao-Liu#7, Fletcher-Powell, and “New function#2”. DE has no problems in optimizing the “New function #1”. But the “New function #2” proves to be a hard nut. However, RPS performs much better for such stochastic functions. When the Fletcher-Powell function is optimized with non-stochastic c vector, DE works fine. But as soon as c is stochastic, it becomes unstable. Thus, it may be observed that an introduction of stochasticity into the decision variables (or simply added to the function as in Yao-Liu#7) interferes with the fundamentals of DE, which works through attainment of better and better (in the sense of Pareto improvement) population at each successive iteration. The paper concludes: (1) for different types of problems, different schemes of crossover (including none) may be suitable or unsuitable, (2) Stochasticity entering into the optimand function may make DE unstable, but RPS may function well.Differential Evolution; Repulsive Particle Swarm; Global optimization; non-convex functions; Fortran; computer program; benchmark; test; Stochastic functions; Fletcher-Powell; Kowalik; Hougen; Power-sum; Perm; Zero-sum; New functions; Bukin function
- …