2,244 research outputs found
Optimal advertising campaign generation for multiple brands using MOGA
The paper proposes a new modified multiobjective
genetic algorithm (MOGA) for the problem of optimal television (TV) advertising campaign generation for multiple brands. This NP-hard combinatorial optimization problem with numerous constraints is one of the key issues for an advertising agency when producing the optimal TV mediaplan. The classical approach to the solution of this problem is the greedy heuristic, which relies on the strength of the preceding commercial breaks when selecting
the next break to add to the campaign. While the greedy heuristic is capable of generating only a group of solutions that are closely related in the objective space, the proposed modified MOGA produces a Pareto-optimal set of chromosomes that: 1) outperform the greedy heuristic and 2) let the mediaplanner choose from a variety of uniformly distributed tradeoff solutions. To achieve these
results, the special problem-specific solution encoding, genetic operators, and original local optimization routine were developed for the algorithm. These techniques allow the algorithm to manipulate with only feasible individuals, thus, significantly improving its performance that is complicated by the problem constraints. The efficiency of the developed optimization method is verified using
the real data sets from the Canadian advertising industry
The Kalai-Smorodinski solution for many-objective Bayesian optimization
An ongoing aim of research in multiobjective Bayesian optimization is to
extend its applicability to a large number of objectives. While coping with a
limited budget of evaluations, recovering the set of optimal compromise
solutions generally requires numerous observations and is less interpretable
since this set tends to grow larger with the number of objectives. We thus
propose to focus on a specific solution originating from game theory, the
Kalai-Smorodinsky solution, which possesses attractive properties. In
particular, it ensures equal marginal gains over all objectives. We further
make it insensitive to a monotonic transformation of the objectives by
considering the objectives in the copula space. A novel tailored algorithm is
proposed to search for the solution, in the form of a Bayesian optimization
algorithm: sequential sampling decisions are made based on acquisition
functions that derive from an instrumental Gaussian process prior. Our approach
is tested on four problems with respectively four, six, eight, and nine
objectives. The method is available in the Rpackage GPGame available on CRAN at
https://cran.r-project.org/package=GPGame
Preference Articulation by Means of the R2 Indicator
International audienceIn multi-objective optimization, set-based performance indicators have become the state of the art for assessing the quality of Pareto front approximations. As a consequence, they are also more and more used within the design of multi-objective optimization algorithms. The R2 and the Hypervolume (HV) indicator represent two popular examples. In order to understand the behavior and the approximations preferred by these indicators and algorithms, a comprehensive knowledge of the indicator's properties is required. Whereas this knowledge is available for the HV, we presented a first approach in this direction for the R2 indicator just recently. In this paper, we build upon this knowledge and enhance the considerations with respect to the integration of preferences into the R2 indicator. More specifically, we analyze the effect of the reference point, the domain of the weights, and the distribution of weight vectors on the optimization of μ solutions with respect to the R2 indicator. By means of theoretical findings and empirical evidence, we show the potentials of these three possibilities using the optimal distribution of μ solutions for exemplary setups
A generic optimising feature extraction method using multiobjective genetic programming
In this paper, we present a generic, optimising feature extraction method using multiobjective genetic programming. We re-examine the feature extraction problem and show that effective feature extraction can significantly enhance the performance of pattern recognition systems with simple classifiers. A framework is presented to evolve optimised feature extractors that transform an input pattern space into a decision space in which maximal class separability is obtained. We have applied this method to real world datasets from the UCI Machine Learning and StatLog databases to verify our approach and compare our proposed method with other reported results. We conclude that our algorithm is able to produce classifiers of superior (or equivalent) performance to the conventional classifiers examined, suggesting removal of the need to exhaustively evaluate a large family of conventional classifiers on any new problem. (C) 2010 Elsevier B.V. All rights reserved
Evolutionary Many-objective Optimization of Hybrid Electric Vehicle Control: From General Optimization to Preference Articulation
Many real-world optimization problems have more than three objectives, which has triggered increasing research interest in developing efficient and effective evolutionary algorithms for solving many-objective optimization problems. However, most many-objective evolutionary algorithms have only been evaluated on benchmark test functions and few applied to real-world optimization problems. To move a step forward, this paper presents a case study of solving a many-objective hybrid electric vehicle controller design problem using three state-of-the-art algorithms, namely, a decomposition based evolutionary algorithm (MOEA/D), a non-dominated sorting based genetic algorithm (NSGA-III), and a reference vector guided evolutionary algorithm (RVEA). We start with a typical setting aiming at approximating the Pareto front without introducing any user preferences. Based on the analyses of the approximated Pareto front, we introduce a preference articulation method and embed it in the three evolutionary algorithms for identifying solutions that the decision-maker prefers. Our experimental results demonstrate that by incorporating user preferences into many-objective evolutionary algorithms, we are not only able to gain deep insight into the trade-off relationships between the objectives, but also to achieve high-quality solutions reflecting the decision-maker’s preferences. In addition, our experimental results indicate that each of the three algorithms examined in this work has its unique advantages that can be exploited when applied to the optimization of real-world problems
Quality Indicators for Preference-based Evolutionary Multi-objective Optimization Using a Reference Point: A Review and Analysis
Some quality indicators have been proposed for benchmarking preference-based
evolutionary multi-objective optimization algorithms using a reference point.
Although a systematic review and analysis of the quality indicators are helpful
for both benchmarking and practical decision-making, neither has been
conducted. In this context, first, this paper reviews existing regions of
interest and quality indicators for preference-based evolutionary
multi-objective optimization using the reference point. We point out that each
quality indicator was designed for a different region of interest. Then, this
paper investigates the properties of the quality indicators. We demonstrate
that an achievement scalarizing function value is not always consistent with
the distance from a solution to the reference point in the objective space. We
observe that the regions of interest can be significantly different depending
on the position of the reference point and the shape of the Pareto front. We
identify undesirable properties of some quality indicators. We also show that
the ranking of preference-based evolutionary multi-objective optimization
algorithms depends on the choice of quality indicators
Which Surrogate Works for Empirical Performance Modelling? A Case Study with Differential Evolution
It is not uncommon that meta-heuristic algorithms contain some intrinsic
parameters, the optimal configuration of which is crucial for achieving their
peak performance. However, evaluating the effectiveness of a configuration is
expensive, as it involves many costly runs of the target algorithm. Perhaps
surprisingly, it is possible to build a cheap-to-evaluate surrogate that models
the algorithm's empirical performance as a function of its parameters. Such
surrogates constitute an important building block for understanding algorithm
performance, algorithm portfolio/selection, and the automatic algorithm
configuration. In principle, many off-the-shelf machine learning techniques can
be used to build surrogates. In this paper, we take the differential evolution
(DE) as the baseline algorithm for proof-of-concept study. Regression models
are trained to model the DE's empirical performance given a parameter
configuration. In particular, we evaluate and compare four popular regression
algorithms both in terms of how well they predict the empirical performance
with respect to a particular parameter configuration, and also how well they
approximate the parameter versus the empirical performance landscapes
- …