909 research outputs found
Evolutionary Multiobjective Optimization Driven by Generative Adversarial Networks (GANs)
Recently, increasing works have proposed to drive evolutionary algorithms
using machine learning models. Usually, the performance of such model based
evolutionary algorithms is highly dependent on the training qualities of the
adopted models. Since it usually requires a certain amount of data (i.e. the
candidate solutions generated by the algorithms) for model training, the
performance deteriorates rapidly with the increase of the problem scales, due
to the curse of dimensionality. To address this issue, we propose a
multi-objective evolutionary algorithm driven by the generative adversarial
networks (GANs). At each generation of the proposed algorithm, the parent
solutions are first classified into real and fake samples to train the GANs;
then the offspring solutions are sampled by the trained GANs. Thanks to the
powerful generative ability of the GANs, our proposed algorithm is capable of
generating promising offspring solutions in high-dimensional decision space
with limited training data. The proposed algorithm is tested on 10 benchmark
problems with up to 200 decision variables. Experimental results on these test
problems demonstrate the effectiveness of the proposed algorithm
Meta-heuristic algorithms in car engine design: a literature survey
Meta-heuristic algorithms are often inspired by natural phenomena, including the evolution of species in Darwinian natural selection theory, ant behaviors in biology, flock behaviors of some birds, and annealing in metallurgy. Due to their great potential in solving difficult optimization problems, meta-heuristic algorithms have found their way into automobile engine design. There are different optimization problems arising in different areas of car engine management including calibration, control system, fault diagnosis, and modeling. In this paper we review the state-of-the-art applications of different meta-heuristic algorithms in engine management systems. The review covers a wide range of research, including the application of meta-heuristic algorithms in engine calibration, optimizing engine control systems, engine fault diagnosis, and optimizing different parts of engines and modeling. The meta-heuristic algorithms reviewed in this paper include evolutionary algorithms, evolution strategy, evolutionary programming, genetic programming, differential evolution, estimation of distribution algorithm, ant colony optimization, particle swarm optimization, memetic algorithms, and artificial immune system
A Recommender System Approach for Very Large-scale Multiobjective Optimization
We define very large multi-objective optimization problems to be
multiobjective optimization problems in which the number of decision variables
is greater than 100,000 dimensions. This is an important class of problems as
many real-world problems require optimizing hundreds of thousands of variables.
Existing evolutionary optimization methods fall short of such requirements when
dealing with problems at this very large scale. Inspired by the success of
existing recommender systems to handle very large-scale items with limited
historical interactions, in this paper we propose a method termed Very
large-scale Multiobjective Optimization through Recommender Systems (VMORS).
The idea of the proposed method is to transform the defined such very
large-scale problems into a problem that can be tackled by a recommender
system. In the framework, the solutions are regarded as users, and the
different evolution directions are items waiting for the recommendation. We use
Thompson sampling to recommend the most suitable items (evolutionary
directions) for different users (solutions), in order to locate the optimal
solution to a multiobjective optimization problem in a very large search space
within acceptable time. We test our proposed method on different problems from
100,000 to 500,000 dimensions, and experimental results show that our method
not only shows good performance but also significant improvement over existing
methods.Comment: 12 pages, 6 figure
Multiple source transfer learning for dynamic multiobjective optimization
Recently, dynamic multiobjective evolutionary algorithms (DMOEAs) with transfer learning have become popular for solving dynamic multiobjective optimization problems (DMOPs), as the used transfer learning methods in DMOEAs can effectively generate a good initial population for the new environment. However, most of them only transfer non-dominated solutions from the previous one or two environments, which cannot fully exploit all historical information and may easily induce negative transfer as only limited knowledge is available. To address this problem, this paper presents a multiple source transfer learning method for DMOEA, called MSTL-DMOEA, which runs two transfer learning procedures to fully exploit the historical information from all previous environments. First, to select some representative solutions for knowledge transfer, one clustering-based manifold transfer learning is run to cluster non-dominated solutions of the last environment to obtain their centroids, which are then fed into the manifold transfer learning model to predict the corresponding centroids for the new environment. After that, multiple source transfer learning is further run by using multisource TrAdaboost, which can fully exploit information from the above centroids in new environment and old centroids from all previous environments, aiming to construct a more accurate prediction model. This way, MSTL-DMOEA can predict an initial population with better quality for the new environment. The experimental results also validate the superiority of MSTL-DMOEA over several competitive state-of-the-art DMOEAs in solving various kinds of DMOPs
Evolutionary Dynamic Multi-Objective Optimisation : A survey
Peer reviewedPostprin
Evolutionary Dynamic Optimization and Machine Learning
Evolutionary Computation (EC) has emerged as a powerful field of Artificial
Intelligence, inspired by nature's mechanisms of gradual development. However,
EC approaches often face challenges such as stagnation, diversity loss,
computational complexity, population initialization, and premature convergence.
To overcome these limitations, researchers have integrated learning algorithms
with evolutionary techniques. This integration harnesses the valuable data
generated by EC algorithms during iterative searches, providing insights into
the search space and population dynamics. Similarly, the relationship between
evolutionary algorithms and Machine Learning (ML) is reciprocal, as EC methods
offer exceptional opportunities for optimizing complex ML tasks characterized
by noisy, inaccurate, and dynamic objective functions. These hybrid techniques,
known as Evolutionary Machine Learning (EML), have been applied at various
stages of the ML process. EC techniques play a vital role in tasks such as data
balancing, feature selection, and model training optimization. Moreover, ML
tasks often require dynamic optimization, for which Evolutionary Dynamic
Optimization (EDO) is valuable. This paper presents the first comprehensive
exploration of reciprocal integration between EDO and ML. The study aims to
stimulate interest in the evolutionary learning community and inspire
innovative contributions in this domain
- …