25 research outputs found

    Regarding the behavior of bison runners within the Bison algorithm

    Get PDF
    This paper proposes a modification of the Bison Algorithm’s running technique, which allows the running group to exploit the areas of discovered promising solutions. It also provides a closer examination of the successful running behavior and its impact on the overall optimization process. The new algorithm is then compared to other optimization algorithms on the IEEE CEC 2017 benchmark solving continuous minimization problems. © 2018, Brno University of Technology. All rights reserved

    How does the number of objective function evaluations impact our understanding of metaheuristics behavior?

    Get PDF
    Comparing various metaheuristics based on an equal number of objective function evaluations has become standard practice. Many contemporary publications use a specific number of objective function evaluations by the benchmarking sets definitions. Furthermore, many publications deal with the recurrent theme of late stagnation, which may lead to the impression that continuing the optimization process could be a waste of computational capabilities. But is it? Recently, many challenges, issues, and questions have been raised regarding fair comparisons and recommendations towards good practices for benchmarking metaheuristic algorithms. The aim of this work is not to compare the performance of several well-known algorithms but to investigate the issues that can appear in benchmarking and comparisons of metaheuristics performance (no matter what the problem is). This article studies the impact of a higher evaluation number on a selection of metaheuristic algorithms. We examine the effect of a raised evaluation budget on overall performance, mean convergence, and population diversity of selected swarm algorithms and IEEE CEC competition winners. Even though the final impact varies based on current algorithm selection, it may significantly affect the final verdict of metaheuristics comparison. This work has picked an important benchmarking issue and made extensive analysis, resulting in conclusions and possible recommendations for users working with real engineering optimization problems or researching the metaheuristics algorithms. Especially nowadays, when metaheuristic algorithms are used for increasingly complex optimization problems, and meet machine learning in AutoML frameworks, we conclude that the objective function evaluation budget should be considered another vital optimization input variable.Internal Grant Agency of Tomas Bata University [IGA/CebiaTech/2021/001]; AI Laboratory, Faculty of Applied Informatics, Tomas Bata University in ZlinIGA/CebiaTech/2021/001; Univerzita Tomáše Bati ve Zlín

    Regarding the Behavior of Bison Runners Within the Bison Algorithm

    Get PDF
    This paper proposes a modification of the Bison Algorithm’s running technique, which allows the running group to exploit the areas of discovered promising solutions. It also provides a closer examination of the successful running behavior and its impact on the overall optimization process. The new algorithm is then compared to other optimization algorithms on the IEEE CEC 2017 benchmark solving continuous minimization problems

    Why tuning the control parameters of metaheuristic algorithms is so important for fair comparison?

    Get PDF
    Although metaheuristic optimization has become a common practice, new bio-inspired algorithms often suffer from a priori ill reputation. One of the reasons is a common bad practice in metaheuristic proposals. It is essential to pay attention to the quality of conducted experiments, especially when comparing several algorithms among themselves. The comparisons should be fair and unbiased. This paper points to the importance of proper initial parameter configurations of the compared algorithms. We highlight the performance differences with several popular and recommended parameter configurations. Even though the parameter selection was mostly based on comprehensive tuning experiments, the algorithms’ performance was surprisingly inconsistent, given various parameter settings. Based on the presented evidence, we conclude that paying attention to the metaheuristic algorithm’s parameter tuning should be an integral part of the development and testing processes. © 2020, Brno University of Technology. All rights reserved

    Relating Training Instances to Automatic Design of Algorithms for Bin Packing via Features (Detailed Experiments and Results)

    Get PDF
    Automatic Design of Algorithms (ADA) shifts the burden of algorithm choice and design from developer to machine. Constructing an appropriate solver from a set of problem instances becomes a machine learning problem, with instances as training data. An efficient solver is trained for unseen problem instances with similar characteristics to those in the training set. However, this paper reveals that, as with classification and regression, for ADA not all training sets are equally valuable. We apply a typical genetic programming ADA approach for bin packing problems to several new and existing public benchmark sets. Algorithms trained on some sets are general and apply well to most others, whereas some training sets result in highly specialised algorithms that do not generalise. We relate these findings to features (simple metrics) of instances. Using instance sets with narrowly-distributed features for training results in highly specialised algorithms, whereas those with well-spread features result in very general algorithms. We show that variance in certain features has a strong correlation with the generality of the trained policies. Our results provide further grounding for recent work using features to predict algorithm performance, and show the suitability of particular instance sets for training in ADA for bin packing. The data sets, including all computed features, the evolved policies, and their performances, and the visualisations for all feature sets, are available from http://hdl.handle.net/11667/108.Work funded by UK EPSRC [grants EP/N002849/1, EP/J017515/1]. Results obtained using the EPSRC funded ARCHIE-WeSt HPC [EPSRC grant EP/K000586/1]

    A novel design approach for 5G massive MIMO and NB-IoT green networks using a hybrid Jaya-differential evolution algorithm

    Get PDF
    Our main objective is to reduce power consumption by responding to the instantaneous bit rate demand by the user for 4th Generation (4G) and 5th Generation (5G) Massive MIMO network configurations. Moreover, we present and address the problem of designing green LTE networks with the Internet of Things (IoT) nodes. We consider the new NarrowBand-IoT (NB-IoT) wireless technology that will emerge in current and future access networks. In this context, we apply emerging evolutionary algorithms in the context of green network design. We investigate three different cases to show the performance of the new proposed algorithm, namely the 4G, 5G Massive MIMO, and the NB-IoT technologies. More specifically, we investigate the Teaching-Learning-Optimization (TLBO), the Jaya algorithm, the self-adaptive differential evolution jDE algorithm, and other hybrid algorithms. We introduce a new hybrid algorithm named Jaya-jDE that uses concepts from both Jaya and jDE algorithms in an effective way. The results show that 5G Massive MIMO networks require about 50% less power consumption than the 4G ones, and the NB-IoT in-band deployment requires about 10% less power than guard-band deployment. Moreover, Jaya-jDE emerges as the best algorithm based on the results

    Comprehensive Taxonomies of Nature- and Bio-inspired Optimization: Inspiration versus Algorithmic Behavior, Critical Analysis and Recommendations

    Full text link
    In recent years, a great variety of nature- and bio-inspired algorithms has been reported in the literature. This algorithmic family simulates different biological processes observed in Nature in order to efficiently address complex optimization problems. In the last years the number of bio-inspired optimization approaches in literature has grown considerably, reaching unprecedented levels that dark the future prospects of this field of research. This paper addresses this problem by proposing two comprehensive, principle-based taxonomies that allow researchers to organize existing and future algorithmic developments into well-defined categories, considering two different criteria: the source of inspiration and the behavior of each algorithm. Using these taxonomies we review more than three hundred publications dealing with nature-inspired and bio-inspired algorithms, and proposals falling within each of these categories are examined, leading to a critical summary of design trends and similarities between them, and the identification of the most similar classical algorithm for each reviewed paper. From our analysis we conclude that a poor relationship is often found between the natural inspiration of an algorithm and its behavior. Furthermore, similarities in terms of behavior between different algorithms are greater than what is claimed in their public disclosure: specifically, we show that more than one-third of the reviewed bio-inspired solvers are versions of classical algorithms. Grounded on the conclusions of our critical analysis, we give several recommendations and points of improvement for better methodological practices in this active and growing research field.Comment: 76 pages, 6 figure

    Assessing hyper-heuristic performance

    Get PDF
    Limited attention has been paid to assessing the generality performance of hyper-heuristics. The performance of hyper-heuristics has been predominately assessed in terms of optimality which is not ideal as the aim of hyper-heuristics is not to be competitive with state of the art approaches but rather to raise the level of generality, i.e. the ability of a technique to produce good results for different problem instances or problems rather than the best results for some instances and poor results for others. Furthermore from existing literature in this area it is evident that different hyper-heuristics aim to achieve different levels of generality and need to be assessed as such. To cater for this the paper firstly presents a new taxonomy of four different levels of generality that can be attained by a hyper-heuristic based on a survey of the literature. The paper then proposes a performance measure to assess the performance of different types of hyper-heuristics at the four levels of generality in terms of generality rather than optimality. Three case studies from the literature are used to demonstrate the application of the generality performance measure. The paper concludes by examining how the generality measure can be combined with measures of other performance criteria, such as optimality, to assess hyper-heuristic performance on more than one criterion

    Mathematical Models and Decomposition Algorithms for Cutting and Packing Problems

    Get PDF
    In this thesis, we provide (or review) new and effective algorithms based on Mixed-Integer Linear Programming (MILP) models and/or decomposition approaches to solve exactly various cutting and packing problems. The first three contributions deal with the classical bin packing and cutting stock problems. First, we propose a survey on the problems, in which we review more than 150 references, implement and computationally test the most common methods used to solve the problems (including branch-and-price, constraint programming (CP) and MILP), and we successfully propose new instances that are difficult to solve in practice. Then, we introduce the BPPLIB, a collection of codes, benchmarks, and links for the two problems. Finally, we study in details the main MILP formulations that have been proposed for the problems, we provide a clear picture of the dominance and equivalence relations that exist among them, and we introduce reflect, a new pseudo-polynomial formulation that achieves state of the art results for both problems and some variants. The following three contributions deal with two-dimensional packing problems. First, we propose a method using Logic based Benders’ decomposition for the orthogonal stock cutting problem and some extensions. We solve the master problem through an MILP model while CP is used to solve the slave problem. Computational experiments on classical benchmarks from the literature show the effectiveness of the proposed approach. Then, we introduce TwoBinGame, a visual application we developed for students to interactively solve two-dimensional packing problems, and analyze the results obtained by 200 students. Finally, we study a complex optimization problem that originates from the packaging industry, which combines cutting and scheduling decisions. For its solution, we propose mathematical models and heuristic algorithms that involve a non-trivial decomposition method. In the last contribution, we study and strengthen various MILP and CP approaches for three project scheduling problems

    From algorithm selection to generation using deep learning

    Get PDF
    Algorithm selection and generation techniques are two methods that can be used to exploit the performance complementarity of different algorithms when applied to large diverse sets of combinatorial problem instances. As there is no single algorithm that dominates all others on all problem instances, algorithm selection automatically selects an algorithm expected to perform best for each problem instance. Meanwhile, algorithm generation refers to combining different algorithms in a manner that allows the resulting method to improve the efficacy of a pool of algorithms. This thesis examines algorithm selection and generation within a single streaming problem domain, that is Bin-Packing, where novel approaches are proposed and evaluated on large problem sets. This research starts with presenting a novel feature-free approach to select the best performing heuristic by capturing the sequential information implicit in a streaming instance and using this as direct input to two Deep Learning (DL) models, Long-Short-Term Memory (LSTM) or Gated Recurrent Unit (GRU), to learn a mapping from an instance to an algorithm. Results obtained using the proposed approach show that the performance of the feature-free selectors significantly outperforms the performance of both the single best solver and the classical feature-based approach using well-known Machine Learning (ML) classifiers when applied to large sets of diverse problem instances. Next, a more radical approach is proposed: bypass algorithm selection altogether by training encoder-decoder LSTM using solutions obtained from a set of algorithms to directly predict a solution from the instance data behaving as an automatically generated algorithm. Experiments conducted on large datasets using problem batches of varying sizes show that the generated algorithm is able to accurately predict solutions, particularly with small batch sizes. Finally, the thesis develops the proposed encoder-decoder approach by introducing a novel neural approach for generating algorithms, in which a neural network acts as an algorithm by generating decisions. Two architectures are evaluated, an encoder-decoder LSTM and a feed-forward Neural Network (NN), and trained using the decisions output from existing algorithms on a large set of instances. Experiments show that the new generated algorithms are capable of solving a subset of instances better than the well-known bin-packing algorithms, and hence they can significantly improve the overall performance when they are added to a pool of algorithms
    corecore