895 research outputs found

    The SOS Platform: Designing, Tuning and Statistically Benchmarking Optimisation Algorithms

    Get PDF
    open access articleWe present Stochastic Optimisation Software (SOS), a Java platform facilitating the algorithmic design process and the evaluation of metaheuristic optimisation algorithms. SOS reduces the burden of coding miscellaneous methods for dealing with several bothersome and time-demanding tasks such as parameter tuning, implementation of comparison algorithms and testbed problems, collecting and processing data to display results, measuring algorithmic overhead, etc. SOS provides numerous off-the-shelf methods including: (1) customised implementations of statistical tests, such as the Wilcoxon rank-sum test and the Holm–Bonferroni procedure, for comparing the performances of optimisation algorithms and automatically generating result tables in PDF and formats; (2) the implementation of an original advanced statistical routine for accurately comparing couples of stochastic optimisation algorithms; (3) the implementation of a novel testbed suite for continuous optimisation, derived from the IEEE CEC 2014 benchmark, allowing for controlled activation of the rotation on each testbed function. Moreover, we briefly comment on the current state of the literature in stochastic optimisation and highlight similarities shared by modern metaheuristics inspired by nature. We argue that the vast majority of these algorithms are simply a reformulation of the same methods and that metaheuristics for optimisation should be simply treated as stochastic processes with less emphasis on the inspiring metaphor behind them

    On the detection of nearly optimal solutions in the context of single-objective space mission design problems

    Get PDF
    When making decisions, having multiple options available for a possible realization of the same project can be advantageous. One way to increase the number of interesting choices is to consider, in addition to the optimal solution x*, also nearly optimal or approximate solutions; these alternative solutions differ from x* and can be in different regions – in the design space – but fulfil certain proximity to its function value f(x*). The scope of this article is the efficient computation and discretization of the set E of e–approximate solutions for scalar optimization problems. To accomplish this task, two strategies to archive and update the data of the search procedure will be suggested and investigated. To make emphasis on data storage efficiency, a way to manage significant and insignificant parameters is also presented. Further on, differential evolution will be used together with the new archivers for the computation of E. Finally, the behaviour of the archiver, as well as the efficiency of the resulting search procedure, will be demonstrated on some academic functions as well as on three models related to space mission design

    A Comparison of Several Heuristic Algorithms for Solving High Dimensional Optimization Problems

    Get PDF
    The number of heuristic optimization algorithms has exploded over the last decade with new methods being proposed constantly. A recent overview of existing heuristic methods has listed over 130 algorithms. The majority of these optimization algorithms have been designed and applied to solve real-parameter function optimization problems, each claiming to be superior to other methods in terms of performance. However, most of these algorithms have been tested on relatively low dimensional problems, i.e., problems involving less than 30 parameters. With the recent emergence of Big Data, the existing optimization methods need to be tested to find those (un)suitable to handle highly dimensional problems. This paper represents an initial step in such direction. Three traditional heuristic algorithms are systematically analyzed and tested in detail for problems involving up to 100 parameters. Genetic algorithms (GA), particle swarm optimization (PSO) and differential evolution (DE) are compared in terms of accuracy and runtime, using several high dimensional standard benchmark functions

    A new hybrid evolutionary algorithm for the treatment of equality constrained MOPs

    Get PDF
    Multi-objective evolutionary algorithms are widely used by researchers and practitioners to solve multi-objective optimization problems (MOPs), since they require minimal assumptions and are capable of computing a finite size approximation of the entire solution set in one run of the algorithm. So far, however, the adequate treatment of equality constraints has played a minor role. Equality constraints are particular since they typically reduce the dimension of the search space, which causes problems for stochastic search algorithms such as evolutionary strategies. In this paper, we show that multi-objective evolutionary algorithms hybridized with continuation-like techniques lead to fast and reliable numerical solvers. For this, we first propose three new problems with different characteristics that are indeed hard to solve by evolutionary algorithms. Next, we develop a variant of NSGA-II with a continuation method. We present numerical results on several equality-constrained MOPs to show that the resulting method is highly competitive to state-of-the-art evolutionary algorithms.Peer ReviewedPostprint (published version

    Multi-objective optimisation on motorized momentum exchange tether for payload orbital transfer

    Get PDF
    The symmetrical motorised momentum exchange tether, is intended to be excited by a continuous torque, so that, it can be applied as an orbital transfer system. The motor drive accelerates the tether, and increases the relative velocity of payloads fitted to each end. In order to access better tether performance, a higher efficiency index needs to be achieved. Meanwhile, the stress in each tether sub-span should stay within the stress limitations. The multi-objective optimisation methods of Genetic Algorithms can be applied for tether performance enhancement. The tether's efficiency index and stress are used as multi-objectives, and the analysis of the resulting Pareto front suggests a set of solutions for the parameters of the motorised momentum exchange tether when used for payload transfer, in order to achieve relative high transfer performance, and safe tether strength

    Hybrid non-dominated sorting genetic algorithm with adaptive operators selection

    Get PDF
    Multiobjective optimization entails minimizing or maximizing multiple objective functions subject to a set of constraints. Many real world applications can be formulated as multi-objective optimization problems (MOPs), which often involve multiple conflicting objectives to be optimized simultaneously. Recently, a number of multi-objective evolutionary algorithms (MOEAs) were developed suggested for these MOPs as they do not require problem specific information. They find a set of non-dominated solutions in a single run. The evolutionary process on which they are based, typically relies on a single genetic operator. Here, we suggest an algorithm which uses a basket of search operators. This is because it is never easy to choose the most suitable operator for a given problem. The novel hybrid non-dominated sorting genetic algorithm (HNSGA) introduced here in this paper and tested on the ZDT (Zitzler-Deb-Thiele) and CEC’09 (2009 IEEE Conference on Evolutionary Computations) benchmark problems specifically formulated for MOEAs. Numerical results prove that the proposed algorithm is competitive with state-of-the-art MOEAs

    A Comparison of Selected Modifications of the Particle Swarm Optimization Algorithm

    Get PDF
    We compare 27 modifications of the original particle swarm optimization (PSO) algorithm. The analysis evaluated nine basic PSO types, which differ according to the swarm evolution as controlled by various inertia weights and constriction factor. Each of the basic PSO modifications was analyzed using three different distributed strategies. In the first strategy, the entire swarm population is considered as one unit (OC-PSO), the second strategy periodically partitions the population into equally large complexes according to the particle’s functional value (SCE-PSO), and the final strategy periodically splits the swarm population into complexes using random permutation (SCERand-PSO). All variants are tested using 11 benchmark functions that were prepared for the special session on real-parameter optimization of CEC 2005. It was found that the best modification of the PSO algorithm is a variant with adaptive inertia weight. The best distribution strategy is SCE-PSO, which gives better results than do OC-PSO and SCERand-PSO for seven functions. The sphere function showed no significant difference between SCE-PSO and SCERand-PSO. It follows that a shuffling mechanism improves the optimization process

    Hybrid adaptive evolutionary algorithm based on decomposition

    Get PDF
    The performance of search operators varies across the different stages of the search/optimization process of evolutionary algorithms (EAs). In general, a single search operator may not do well in all these stages when dealing with different optimization and search problems. To mitigate this, adaptive search operator schemes have been introduced. The idea is that when a search operator hits a difficult patch (under-performs) in the search space, the EA scheme “reacts” to that by potentially calling upon a different search operator. Hence, several multiple-search operator schemes have been proposed and employed within EA. In this paper, a hybrid adaptive evolutionary algorithm based on decomposition (HAEA/D) that employs four different crossover operators is suggested. Its performance has been evaluated on the well-known IEEE CEC’09 test instances. HAEA/D has generated promising results which compare well against several well-known algorithms including MOEA/D, on a number of metrics such as the inverted generational distance (IGD), the hyper-volume, the Gamma and Delta functions. These results are included and discussed in this paper

    A Classification of Hyper-heuristic Approaches

    Get PDF
    The current state of the art in hyper-heuristic research comprises a set of approaches that share the common goal of automating the design and adaptation of heuristic methods to solve hard computational search problems. The main goal is to produce more generally applicable search methodologies. In this chapter we present and overview of previous categorisations of hyper-heuristics and provide a unified classification and definition which captures the work that is being undertaken in this field. We distinguish between two main hyper-heuristic categories: heuristic selection and heuristic generation. Some representative examples of each category are discussed in detail. Our goal is to both clarify the main features of existing techniques and to suggest new directions for hyper-heuristic research
    • 

    corecore