15,668 research outputs found
Evolutionary Multiobjective Optimization Driven by Generative Adversarial Networks (GANs)
Recently, increasing works have proposed to drive evolutionary algorithms
using machine learning models. Usually, the performance of such model based
evolutionary algorithms is highly dependent on the training qualities of the
adopted models. Since it usually requires a certain amount of data (i.e. the
candidate solutions generated by the algorithms) for model training, the
performance deteriorates rapidly with the increase of the problem scales, due
to the curse of dimensionality. To address this issue, we propose a
multi-objective evolutionary algorithm driven by the generative adversarial
networks (GANs). At each generation of the proposed algorithm, the parent
solutions are first classified into real and fake samples to train the GANs;
then the offspring solutions are sampled by the trained GANs. Thanks to the
powerful generative ability of the GANs, our proposed algorithm is capable of
generating promising offspring solutions in high-dimensional decision space
with limited training data. The proposed algorithm is tested on 10 benchmark
problems with up to 200 decision variables. Experimental results on these test
problems demonstrate the effectiveness of the proposed algorithm
On testing global optimization algorithms for space trajectory design
In this paper we discuss the procedures to test a global search algorithm applied to a space trajectory design problem. Then, we present some performance indexes that can
be used to evaluate the effectiveness of global optimization algorithms. The performance
indexes are then compared highlighting the actual significance of each one of them. A number of global optimization algorithms are tested on four typical space trajectory design problems. From the results of the proposed testing procedure we infer for each pair
algorithm-problem the relation between the heuristics implemented in the solution algorithm and the main characteristics of the problem under investigation. From this analysis we derive a novel interpretation of some evolutionary heuristics, based on dynamical system
theory and we significantly improve the performance of one of the tested algorithms
A simplex-like search method for bi-objective optimization
We describe a new algorithm for bi-objective optimization, similar to the Nelder Mead simplex
algorithm, widely used for single objective optimization. For diferentiable bi-objective functions on
a continuous search space, internal Pareto optima occur where the two gradient vectors point in
opposite directions. So such optima may be located by minimizing the cosine of the angle between
these vectors. This requires a complex rather than a simplex, so we term the technique the \cosine
seeking complex". An extra beneft of this approach is that a successful search identifes the direction
of the effcient curve of Pareto points, expediting further searches. Results are presented for some
standard test functions. The method presented is quite complicated and space considerations here
preclude complete details. We hope to publish a fuller description in another place
Recommended from our members
Combinatorial optimization and metaheuristics
Today, combinatorial optimization is one of the youngest and most active areas of discrete mathematics. It is a branch of optimization in applied mathematics and computer science, related to operational research, algorithm theory and computational complexity theory. It sits at the intersection of several fields, including artificial intelligence, mathematics and software engineering. Its increasing interest arises for the fact that a large number of scientific and industrial problems can be formulated as abstract combinatorial optimization problems, through graphs and/or (integer) linear programs. Some of these problems have polynomial-time (“efficient”) algorithms, while most of them are NP-hard, i.e. it is not proved that they can be solved in polynomial-time. Mainly, it means that it is not possible to guarantee that an exact solution to the problem can be found and one has to settle for an approximate solution with known performance guarantees. Indeed, the goal of approximate methods is to find “quickly” (reasonable run-times), with “high” probability, provable “good” solutions (low error from the real optimal solution). In the last 20 years, a new kind of algorithm commonly called metaheuristics have emerged in this class, which basically try to combine heuristics in high level frameworks aimed at efficiently and effectively exploring the search space. This report briefly outlines the components, concepts, advantages and disadvantages of different metaheuristic approaches from a conceptual point of view, in order to analyze their similarities and differences. The two very significant forces of intensification and diversification, that mainly determine the behavior of a metaheuristic, will be pointed out. The report concludes by exploring the importance of hybridization and integration methods
Optimisation of Mobile Communication Networks - OMCO NET
The mini conference “Optimisation of Mobile Communication Networks” focuses on advanced methods for search and optimisation applied to wireless communication networks. It is sponsored by Research & Enterprise Fund Southampton Solent University.
The conference strives to widen knowledge on advanced search methods capable of optimisation of wireless communications networks. The aim is to provide a forum for exchange of recent knowledge, new ideas and trends in this progressive and challenging area. The conference will popularise new successful approaches on resolving hard tasks such as minimisation of transmit power, cooperative and optimal routing
- …