12,094 research outputs found
TOWARDS A UNIFIED VIEW OF METAHEURISTICS
This talk provides a complete background on metaheuristics and presents in a unified view the main design questions for all families of metaheuristics and clearly illustrates how to implement the algorithms under a software framework to reuse both the design and code. The key search components of metaheuristics are considered as a toolbox for:
- Designing efficient metaheuristics (e.g. local search, tabu search, simulated annealing, evolutionary algorithms, particle swarm optimization, scatter search, ant colonies, bee colonies, artificial immune systems) for optimization problems.
- Designing efficient metaheuristics for multi-objective optimization problems.
- Designing hybrid, parallel and distributed metaheuristics.
- Implementing metaheuristics on sequential and parallel machines
Review of Metaheuristics and Generalized Evolutionary Walk Algorithm
Metaheuristic algorithms are often nature-inspired, and they are becoming
very powerful in solving global optimization problems. More than a dozen of
major metaheuristic algorithms have been developed over the last three decades,
and there exist even more variants and hybrid of metaheuristics. This paper
intends to provide an overview of nature-inspired metaheuristic algorithms,
from a brief history to their applications. We try to analyze the main
components of these algorithms and how and why they works. Then, we intend to
provide a unified view of metaheuristics by proposing a generalized
evolutionary walk algorithm (GEWA). Finally, we discuss some of the important
open questions.Comment: 14 page
Metaheuristic design of feedforward neural networks: a review of two decades of research
Over the past two decades, the feedforward neural network (FNN) optimization has been a key interest among the researchers and practitioners of multiple disciplines. The FNN optimization is often viewed from the various perspectives: the optimization of weights, network architecture, activation nodes, learning parameters, learning environment, etc. Researchers adopted such different viewpoints mainly to improve the FNN's generalization ability. The gradient-descent algorithm such as backpropagation has been widely applied to optimize the FNNs. Its success is evident from the FNN's application to numerous real-world problems. However, due to the limitations of the gradient-based optimization methods, the metaheuristic algorithms including the evolutionary algorithms, swarm intelligence, etc., are still being widely explored by the researchers aiming to obtain generalized FNN for a given problem. This article attempts to summarize a broad spectrum of FNN optimization methodologies including conventional and metaheuristic approaches. This article also tries to connect various research directions emerged out of the FNN optimization practices, such as evolving neural network (NN), cooperative coevolution NN, complex-valued NN, deep learning, extreme learning machine, quantum NN, etc. Additionally, it provides interesting research challenges for future research to cope-up with the present information processing era
Recommended from our members
Solving the minimum labelling spanning tree problem using hybrid local search
Given a connected, undirected graph whose edges are labelled (or coloured), the minimum
labelling spanning tree (MLST) problem seeks a spanning tree whose edges have the smallest
number of distinct labels (or colours). In recent work, the MLST problem has been shown
to be NP-hard and some effective heuristics (Modified Genetic Algorithm (MGA) and Pilot
Method (PILOT)) have been proposed and analyzed. A hybrid local search method, that we
call Group-Swap Variable Neighbourhood Search (GS-VNS), is proposed in this paper. It is
obtained by combining two classic metaheuristics: Variable Neighbourhood Search (VNS) and
Simulated Annealing (SA). Computational experiments show that GS-VNS outperforms MGA
and PILOT. Furthermore, a comparison with the results provided by an exact approach shows
that we may quickly obtain optimal or near-optimal solutions with the proposed heuristic
Non-linear great deluge with learning mechanism for solving the course timetabling problem
International audienc
Recommended from our members
Combinatorial optimization and metaheuristics
Today, combinatorial optimization is one of the youngest and most active areas of discrete mathematics. It is a branch of optimization in applied mathematics and computer science, related to operational research, algorithm theory and computational complexity theory. It sits at the intersection of several fields, including artificial intelligence, mathematics and software engineering. Its increasing interest arises for the fact that a large number of scientific and industrial problems can be formulated as abstract combinatorial optimization problems, through graphs and/or (integer) linear programs. Some of these problems have polynomial-time (“efficient”) algorithms, while most of them are NP-hard, i.e. it is not proved that they can be solved in polynomial-time. Mainly, it means that it is not possible to guarantee that an exact solution to the problem can be found and one has to settle for an approximate solution with known performance guarantees. Indeed, the goal of approximate methods is to find “quickly” (reasonable run-times), with “high” probability, provable “good” solutions (low error from the real optimal solution). In the last 20 years, a new kind of algorithm commonly called metaheuristics have emerged in this class, which basically try to combine heuristics in high level frameworks aimed at efficiently and effectively exploring the search space. This report briefly outlines the components, concepts, advantages and disadvantages of different metaheuristic approaches from a conceptual point of view, in order to analyze their similarities and differences. The two very significant forces of intensification and diversification, that mainly determine the behavior of a metaheuristic, will be pointed out. The report concludes by exploring the importance of hybridization and integration methods
An Efficient Hybrid Ant Colony System for the Generalized Traveling Salesman Problem
The Generalized Traveling Salesman Problem (GTSP) is an extension of the
well-known Traveling Salesman Problem (TSP), where the node set is partitioned
into clusters, and the objective is to find the shortest cycle visiting each
cluster exactly once. In this paper, we present a new hybrid Ant Colony System
(ACS) algorithm for the symmetric GTSP. The proposed algorithm is a
modification of a simple ACS for the TSP improved by an efficient GTSP-specific
local search procedure. Our extensive computational experiments show that the
use of the local search procedure dramatically improves the performance of the
ACS algorithm, making it one of the most successful GTSP metaheuristics to
date.Comment: 7 page
A statistical learning based approach for parameter fine-tuning of metaheuristics
Metaheuristics are approximation methods used to solve combinatorial optimization problems. Their performance usually depends on a set of parameters that need to be adjusted. The selection of appropriate parameter values causes a loss of efficiency, as it requires time, and advanced analytical and problem-specific skills. This paper provides an overview of the principal approaches to tackle the Parameter Setting Problem, focusing on the statistical procedures employed so far by the scientific community. In addition, a novel methodology is proposed, which is tested using an already existing algorithm for solving the Multi-Depot Vehicle Routing Problem.Peer ReviewedPostprint (published version
- …