7,784 research outputs found
Evolutionary Computation in High Energy Physics
Evolutionary Computation is a branch of computer science with which,
traditionally, High Energy Physics has fewer connections. Its methods were
investigated in this field, mainly for data analysis tasks. These methods and
studies are, however, less known in the high energy physics community and this
motivated us to prepare this lecture. The lecture presents a general overview
of the main types of algorithms based on Evolutionary Computation, as well as a
review of their applications in High Energy Physics.Comment: Lecture presented at 2006 Inverted CERN School of Computing; to be
published in the school proceedings (CERN Yellow Report
Constructing multiple unique input/output sequences using metaheuristic optimisation techniques
Multiple unique input/output sequences (UIOs) are often used to generate robust and compact test sequences in finite state machine (FSM) based testing. However, computing UIOs is NP-hard. Metaheuristic optimisation techniques (MOTs) such as genetic algorithms (GAs) and simulated annealing (SA) are effective in providing good solutions for some NP-hard problems. In the paper, the authors investigate the construction of UIOs by using MOTs. They define a fitness function to guide the search for potential UIOs and use sharing techniques to encourage MOTs to locate UIOs that are calculated as local optima in a search domain. They also compare the performance of GA and SA for UIO construction. Experimental results suggest that, after using a sharing technique, both GA and SA can find a majority of UIOs from the models under test
A hierarchical Mamdani-type fuzzy modelling approach with new training data selection and multi-objective optimisation mechanisms: A special application for the prediction of mechanical properties of alloy steels
In this paper, a systematic data-driven fuzzy modelling methodology is proposed, which allows to construct Mamdani fuzzy models considering both accuracy (precision) and transparency (interpretability) of fuzzy systems. The new methodology employs a fast hierarchical clustering algorithm to generate an initial fuzzy model efficiently; a training data selection mechanism is developed to identify appropriate and efficient data as learning samples; a high-performance Particle Swarm Optimisation (PSO) based multi-objective optimisation mechanism is developed to further improve the fuzzy model in terms of both the structure and the parameters; and a new tolerance analysis method is proposed to derive the confidence bands relating to the final elicited models. This proposed modelling approach is evaluated using two benchmark problems and is shown to outperform other modelling approaches. Furthermore, the proposed approach is successfully applied to complex high-dimensional modelling problems for manufacturing of alloy steels, using ‘real’ industrial data. These problems concern the prediction of the mechanical properties of alloy steels by correlating them with the heat treatment process conditions as well as the weight percentages of the chemical compositions
Composing Distributed Data-intensive Web Services Using a Flexible Memetic Algorithm
Web Service Composition (WSC) is a particularly promising application of Web
services, where multiple individual services with specific functionalities are
composed to accomplish a more complex task, which must fulfil functional
requirements and optimise Quality of Service (QoS) attributes, simultaneously.
Additionally, large quantities of data, produced by technological advances,
need to be exchanged between services. Data-intensive Web services, which
manipulate and deal with those data, are of great interest to implement
data-intensive processes, such as distributed Data-intensive Web Service
Composition (DWSC). Researchers have proposed Evolutionary Computing (EC)
fully-automated WSC techniques that meet all the above factors. Some of these
works employed Memetic Algorithms (MAs) to enhance the performance of EC
through increasing its exploitation ability of in searching neighbourhood area
of a solution. However, those works are not efficient or effective. This paper
proposes an MA-based approach to solving the problem of distributed DWSC in an
effective and efficient manner. In particular, we develop an MA that hybridises
EC with a flexible local search technique incorporating distance of services.
An evaluation using benchmark datasets is carried out, comparing existing
state-of-the-art methods. Results show that our proposed method has the highest
quality and an acceptable execution time overall.Comment: arXiv admin note: text overlap with arXiv:1901.0556
Evolutionary improvement of programs
Most applications of genetic programming (GP) involve the creation of an entirely new function, program or expression to solve a specific problem. In this paper, we propose a new approach that applies GP to improve existing software by optimizing its non-functional properties such as execution time, memory usage, or power consumption. In general, satisfying non-functional requirements is a difficult task and often achieved in part by optimizing compilers. However, modern compilers are in general not always able to produce semantically equivalent alternatives that optimize non-functional properties, even if such alternatives are known to exist: this is usually due to the limited local nature of such optimizations. In this paper, we discuss how best to combine and extend the existing evolutionary methods of GP, multiobjective optimization, and coevolution in order to improve existing software. Given as input the implementation of a function, we attempt to evolve a semantically equivalent version, in this case optimized to reduce execution time subject to a given probability distribution of inputs. We demonstrate that our framework is able to produce non-obvious optimizations that compilers are not yet able to generate on eight example functions. We employ a coevolved population of test cases to encourage the preservation of the function's semantics. We exploit the original program both through seeding of the population in order to focus the search, and as an oracle for testing purposes. As well as discussing the issues that arise when attempting to improve software, we employ rigorous experimental method to provide interesting and practical insights to suggest how to address these issues
An Order-based Algorithm for Minimum Dominating Set with Application in Graph Mining
Dominating set is a set of vertices of a graph such that all other vertices
have a neighbour in the dominating set. We propose a new order-based randomised
local search (RLS) algorithm to solve minimum dominating set problem in
large graphs. Experimental evaluation is presented for multiple types of
problem instances. These instances include unit disk graphs, which represent a
model of wireless networks, random scale-free networks, as well as samples from
two social networks and real-world graphs studied in network science. Our
experiments indicate that RLS performs better than both a classical greedy
approximation algorithm and two metaheuristic algorithms based on ant colony
optimisation and local search. The order-based algorithm is able to find small
dominating sets for graphs with tens of thousands of vertices. In addition, we
propose a multi-start variant of RLS that is suitable for solving the
minimum weight dominating set problem. The application of RLS in graph
mining is also briefly demonstrated
Population-based incremental learning with memory scheme for changing environments
Copyright @ 2005 ACMIn recent years there has been a growing interest in studying evolutionary algorithms for dynamic optimization problems due to its importance in real world applications. Several approaches have been developed, such as the memory scheme. This paper investigates the application of the memory scheme for population-based incremental learning (PBIL) algorithms, a class of evolutionary algorithms, for dynamic optimization problems. A PBIL-specific memory scheme is proposed to improve its adaptability in dynamic environments. In this memory scheme the working probability vector is stored together with the best sample it creates in the memory and is used to reactivate old environments when change occurs. Experimental study based on a series of dynamic environments shows the efficiency of the memory scheme for PBILs in dynamic environments. In this paper, the relationship between the memory scheme and the multipopulation scheme for PBILs in dynamic environments is also investigated. The experimental results indicate a negative interaction of the multi-population scheme on the memory scheme for PBILs in the dynamic test environments
- …