6,730 research outputs found
Learning heuristic selection using a time delay neural network for open vehicle routing
A selection hyper-heuristic is a search method that controls a prefixed set of low-level heuristics for solving a given computationally difficult problem. This study investigates a learning-via demonstrations approach generating a selection hyper-heuristic for Open Vehicle Routing Problem (OVRP). As a chosen āexpertā hyper-heuristic is run on a small set of training problem instances, data is collected to learn from the expert regarding how to decide which low-level heuristic to select and apply to the solution in hand during the search process. In this study, a Time Delay Neural Network (TDNN) is used to extract hidden patterns within the collected data in the form of a classifier ,i.e an āapprenticeā hyper-heuristic, which is then used to solve the āunseenā problem instances. Firstly, the parameters of TDNN are tuned using Taguchi orthogonal array as a design of experiments method. Then the influence of extending and enriching the information collected from the expert and fed into TDNN is explored on the behaviour of the generated apprentice hyper-heuristic. The empirical results show that the use of distance between solutions as an additional information collected from the expert generates an apprentice which outperforms the expert algorithm on a benchmark of OVRP instances
A Novel Efficient Task-Assign Route Planning Method for AUV Guidance in a Dynamic Cluttered Environment
Promoting the levels of autonomy facilitates the vehicle in performing
long-range operations with minimum supervision. The capability of Autonomous
Underwater Vehicles (AUVs) to fulfill the mission objectives is directly
influenced by route planning and task assignment system performance.
The system fives the error of "Bad character(s) in field Abstract" for no
reason. Please refer to manuscript for the full abstractComment: 7 pages, 8 figures, conference paper, IEEE Congress on Evolutionary
Computation (CEC). Vancouver, Canada. July 201
Improving performance of a hyper-heuristic using a multilayer perceptron for vehicle routing
A hyper-heuristic is a heuristic optimisation method which generates or selects heuristics (move operators) based on a set of components while solving a computationally difficult problem. Apprenticeship learning arises while observing the behavior of an expert in action. In this study, we use a multilayer perceptron (MLP) as an apprenticeship learning algorithm to improve upon the performance of a state-of-the-art selection hyper-heuristic used as an expert, which was the winner of a cross-domain heuristic search challenge (CHeSC 2011). We collect data based on the relevant actions of the expert while solving selected vehicle routing problem instances from CHeSC 2011. Then an MLP is trained using this data to build a selection hyper-heuristic consisting of a number classifiers for heuristic selection, parameter control, and move-acceptance. The generated selection hyper-heuristic is tested on the unseen vehicle routing problem instances. The empirical results indicate the success of MLP-based hyper-heuristic achieving a better performance than the expert and some previously proposed algorithms
Optimisation of Mobile Communication Networks - OMCO NET
The mini conference āOptimisation of Mobile Communication Networksā focuses on advanced methods for search and optimisation applied to wireless communication networks. It is sponsored by Research & Enterprise Fund Southampton Solent University.
The conference strives to widen knowledge on advanced search methods capable of optimisation of wireless communications networks. The aim is to provide a forum for exchange of recent knowledge, new ideas and trends in this progressive and challenging area. The conference will popularise new successful approaches on resolving hard tasks such as minimisation of transmit power, cooperative and optimal routing
Reinforcement Learning-assisted Evolutionary Algorithm: A Survey and Research Opportunities
Evolutionary algorithms (EA), a class of stochastic search methods based on
the principles of natural evolution, have received widespread acclaim for their
exceptional performance in various real-world optimization problems. While
researchers worldwide have proposed a wide variety of EAs, certain limitations
remain, such as slow convergence speed and poor generalization capabilities.
Consequently, numerous scholars actively explore improvements to algorithmic
structures, operators, search patterns, etc., to enhance their optimization
performance. Reinforcement learning (RL) integrated as a component in the EA
framework has demonstrated superior performance in recent years. This paper
presents a comprehensive survey on integrating reinforcement learning into the
evolutionary algorithm, referred to as reinforcement learning-assisted
evolutionary algorithm (RL-EA). We begin with the conceptual outlines of
reinforcement learning and the evolutionary algorithm. We then provide a
taxonomy of RL-EA. Subsequently, we discuss the RL-EA integration method, the
RL-assisted strategy adopted by RL-EA, and its applications according to the
existing literature. The RL-assisted procedure is divided according to the
implemented functions including solution generation, learnable objective
function, algorithm/operator/sub-population selection, parameter adaptation,
and other strategies. Finally, we analyze potential directions for future
research. This survey serves as a rich resource for researchers interested in
RL-EA as it overviews the current state-of-the-art and highlights the
associated challenges. By leveraging this survey, readers can swiftly gain
insights into RL-EA to develop efficient algorithms, thereby fostering further
advancements in this emerging field.Comment: 26 pages, 16 figure
- ā¦