531 research outputs found
Efficient Computation of Expected Hypervolume Improvement Using Box Decomposition Algorithms
In the field of multi-objective optimization algorithms, multi-objective
Bayesian Global Optimization (MOBGO) is an important branch, in addition to
evolutionary multi-objective optimization algorithms (EMOAs). MOBGO utilizes
Gaussian Process models learned from previous objective function evaluations to
decide the next evaluation site by maximizing or minimizing an infill
criterion. A common criterion in MOBGO is the Expected Hypervolume Improvement
(EHVI), which shows a good performance on a wide range of problems, with
respect to exploration and exploitation. However, so far it has been a
challenge to calculate exact EHVI values efficiently. In this paper, an
efficient algorithm for the computation of the exact EHVI for a generic case is
proposed. This efficient algorithm is based on partitioning the integration
volume into a set of axis-parallel slices. Theoretically, the upper bound time
complexities are improved from previously and , for two- and
three-objective problems respectively, to , which is
asymptotically optimal. This article generalizes the scheme in higher
dimensional case by utilizing a new hyperbox decomposition technique, which was
proposed by D{\"a}chert et al, EJOR, 2017. It also utilizes a generalization of
the multilayered integration scheme that scales linearly in the number of
hyperboxes of the decomposition. The speed comparison shows that the proposed
algorithm in this paper significantly reduces computation time. Finally, this
decomposition technique is applied in the calculation of the Probability of
Improvement (PoI)
Scalarizing Functions in Bayesian Multiobjective Optimization
Scalarizing functions have been widely used to convert a multiobjective
optimization problem into a single objective optimization problem. However,
their use in solving (computationally) expensive multi- and many-objective
optimization problems in Bayesian multiobjective optimization is scarce.
Scalarizing functions can play a crucial role on the quality and number of
evaluations required when doing the optimization. In this article, we study and
review 15 different scalarizing functions in the framework of Bayesian
multiobjective optimization and build Gaussian process models (as surrogates,
metamodels or emulators) on them. We use expected improvement as infill
criterion (or acquisition function) to update the models. In particular, we
compare different scalarizing functions and analyze their performance on
several benchmark problems with different number of objectives to be optimized.
The review and experiments on different functions provide useful insights when
using and selecting a scalarizing function when using a Bayesian multiobjective
optimization method
Fast calculation of multiobjective probability of improvement and expected improvement criteria for Pareto optimization
The use of surrogate based optimization (SBO) is widely spread in engineering design to reduce the number of computational expensive simulations. However, "real-world" problems often consist of multiple, conflicting objectives leading to a set of competitive solutions (the Pareto front). The objectives are often aggregated into a single cost function to reduce the computational cost, though a better approach is to use multiobjective optimization methods to directly identify a set of Pareto-optimal solutions, which can be used by the designer to make more efficient design decisions (instead of weighting and aggregating the costs upfront). Most of the work in multiobjective optimization is focused on multiobjective evolutionary algorithms (MOEAs). While MOEAs are well-suited to handle large, intractable design spaces, they typically require thousands of expensive simulations, which is prohibitively expensive for the problems under study. Therefore, the use of surrogate models in multiobjective optimization, denoted as multiobjective surrogate-based optimization, may prove to be even more worthwhile than SBO methods to expedite the optimization of computational expensive systems. In this paper, the authors propose the efficient multiobjective optimization (EMO) algorithm which uses Kriging models and multiobjective versions of the probability of improvement and expected improvement criteria to identify the Pareto front with a minimal number of expensive simulations. The EMO algorithm is applied on multiple standard benchmark problems and compared against the well-known NSGA-II, SPEA2 and SMS-EMOA multiobjective optimization methods
The Kalai-Smorodinski solution for many-objective Bayesian optimization
An ongoing aim of research in multiobjective Bayesian optimization is to
extend its applicability to a large number of objectives. While coping with a
limited budget of evaluations, recovering the set of optimal compromise
solutions generally requires numerous observations and is less interpretable
since this set tends to grow larger with the number of objectives. We thus
propose to focus on a specific solution originating from game theory, the
Kalai-Smorodinsky solution, which possesses attractive properties. In
particular, it ensures equal marginal gains over all objectives. We further
make it insensitive to a monotonic transformation of the objectives by
considering the objectives in the copula space. A novel tailored algorithm is
proposed to search for the solution, in the form of a Bayesian optimization
algorithm: sequential sampling decisions are made based on acquisition
functions that derive from an instrumental Gaussian process prior. Our approach
is tested on four problems with respectively four, six, eight, and nine
objectives. The method is available in the Rpackage GPGame available on CRAN at
https://cran.r-project.org/package=GPGame
Hypervolume-based Multi-objective Bayesian Optimization with Student-t Processes
Student- processes have recently been proposed as an appealing alternative
non-parameteric function prior. They feature enhanced flexibility and
predictive variance. In this work the use of Student- processes are explored
for multi-objective Bayesian optimization. In particular, an analytical
expression for the hypervolume-based probability of improvement is developed
for independent Student- process priors of the objectives. Its effectiveness
is shown on a multi-objective optimization problem which is known to be
difficult with traditional Gaussian processes.Comment: 5 pages, 3 figure
A Bayesian approach to constrained single- and multi-objective optimization
This article addresses the problem of derivative-free (single- or
multi-objective) optimization subject to multiple inequality constraints. Both
the objective and constraint functions are assumed to be smooth, non-linear and
expensive to evaluate. As a consequence, the number of evaluations that can be
used to carry out the optimization is very limited, as in complex industrial
design optimization problems. The method we propose to overcome this difficulty
has its roots in both the Bayesian and the multi-objective optimization
literatures. More specifically, an extended domination rule is used to handle
objectives and constraints in a unified way, and a corresponding expected
hyper-volume improvement sampling criterion is proposed. This new criterion is
naturally adapted to the search of a feasible point when none is available, and
reduces to existing Bayesian sampling criteria---the classical Expected
Improvement (EI) criterion and some of its constrained/multi-objective
extensions---as soon as at least one feasible point is available. The
calculation and optimization of the criterion are performed using Sequential
Monte Carlo techniques. In particular, an algorithm similar to the subset
simulation method, which is well known in the field of structural reliability,
is used to estimate the criterion. The method, which we call BMOO (for Bayesian
Multi-Objective Optimization), is compared to state-of-the-art algorithms for
single- and multi-objective constrained optimization
A survey on handling computationally expensive multiobjective optimization problems with evolutionary algorithms
This is the author accepted manuscript. The final version is available from Springer Verlag via the DOI in this record.Evolutionary algorithms are widely used for solving multiobjective optimization problems but are often criticized because of a large number of function evaluations needed. Approximations, especially function approximations, also referred to as surrogates or metamodels are commonly used in the literature to reduce the computation time. This paper presents a survey of 45 different recent algorithms proposed in the literature between 2008 and 2016 to handle computationally expensive multiobjective optimization problems. Several algorithms are discussed based on what kind of an approximation such as problem, function or fitness approximation they use. Most emphasis is given to function approximation-based algorithms. We also compare these algorithms based on different criteria such as metamodeling technique and evolutionary algorithm used, type and dimensions of the problem solved, handling constraints, training time and the type of evolution control. Furthermore, we identify and discuss some promising elements and major issues among algorithms in the literature related to using an approximation and numerical settings used. In addition, we discuss selecting an algorithm to solve a given computationally expensive multiobjective optimization problem based on the dimensions in both objective and decision spaces and the computation budget available.The research of Tinkle Chugh was funded by the COMAS Doctoral Program (at the University of JyvÀskylÀ) and FiDiPro Project DeCoMo (funded by Tekes, the Finnish Funding Agency for Innovation), and the research of Dr. Karthik Sindhya was funded by SIMPRO project funded by Tekes as well as DeCoMo
- âŠ