24,982 research outputs found

    A Bayesian approach to constrained single- and multi-objective optimization

    Get PDF
    This article addresses the problem of derivative-free (single- or multi-objective) optimization subject to multiple inequality constraints. Both the objective and constraint functions are assumed to be smooth, non-linear and expensive to evaluate. As a consequence, the number of evaluations that can be used to carry out the optimization is very limited, as in complex industrial design optimization problems. The method we propose to overcome this difficulty has its roots in both the Bayesian and the multi-objective optimization literatures. More specifically, an extended domination rule is used to handle objectives and constraints in a unified way, and a corresponding expected hyper-volume improvement sampling criterion is proposed. This new criterion is naturally adapted to the search of a feasible point when none is available, and reduces to existing Bayesian sampling criteria---the classical Expected Improvement (EI) criterion and some of its constrained/multi-objective extensions---as soon as at least one feasible point is available. The calculation and optimization of the criterion are performed using Sequential Monte Carlo techniques. In particular, an algorithm similar to the subset simulation method, which is well known in the field of structural reliability, is used to estimate the criterion. The method, which we call BMOO (for Bayesian Multi-Objective Optimization), is compared to state-of-the-art algorithms for single- and multi-objective constrained optimization

    Index Information Algorithm with Local Tuning for Solving Multidimensional Global Optimization Problems with Multiextremal Constraints

    Full text link
    Multidimensional optimization problems where the objective function and the constraints are multiextremal non-differentiable Lipschitz functions (with unknown Lipschitz constants) and the feasible region is a finite collection of robust nonconvex subregions are considered. Both the objective function and the constraints may be partially defined. To solve such problems an algorithm is proposed, that uses Peano space-filling curves and the index scheme to reduce the original problem to a H\"{o}lder one-dimensional one. Local tuning on the behaviour of the objective function and constraints is used during the work of the global optimization procedure in order to accelerate the search. The method neither uses penalty coefficients nor additional variables. Convergence conditions are established. Numerical experiments confirm the good performance of the technique.Comment: 29 pages, 5 figure

    A discrete dynamic convexized method for nonlinear integer programming

    Get PDF
    AbstractIn this paper, we consider the box constrained nonlinear integer programming problem. We present an auxiliary function, which has the same discrete global minimizers as the problem. The minimization of the function using a discrete local search method can escape successfully from previously converged discrete local minimizers by taking increasing values of a parameter. We propose an algorithm to find a global minimizer of the box constrained nonlinear integer programming problem. The algorithm minimizes the auxiliary function from random initial points. We prove that the algorithm can converge asymptotically with probability one. Numerical experiments on a set of test problems show that the algorithm is efficient and robust

    A hybrid approach to constrained global optimization

    Get PDF
    In this paper, we propose a novel hybrid global optimization method to solve constrained optimization problems. An exact penalty function is first applied to approximate the original constrained optimization problem by a sequence of optimization problems with bound constraints. To solve each of these box constrained optimization problems, two hybrid methods are introduced, where two different strategies are used to combine limited memory BFGS (L-BFGS) with Greedy Diffusion Search (GDS). The convergence issue of the two hybrid methods is addressed. To evaluate the effectiveness of the proposed algorithm, 18 box constrained and 4 general constrained problems from the literature are tested. Numerical results obtained show that our proposed hybrid algorithm is more effective in obtaining more accurate solutions than those compared to

    Global algorithms for nonlinear discrete optimization and discrete-valued optimal control problems

    Get PDF
    Optimal control problems arise in many applications, such as in economics, finance, process engineering, and robotics. Some optimal control problems involve a control which takes values from a discrete set. These problems are known as discrete-valued optimal control problems. Most practical discrete-valued optimal control problems have multiple local minima and thus require global optimization methods to generate practically useful solutions. Due to the high complexity of these problems, metaheuristic based global optimization techniques are usually required.One of the more recent global optimization tools in the area of discrete optimization is known as the discrete filled function method. The basic idea of the discrete filled function method is as follows. We choose an initial point and then perform a local search to find an initial local minimizer. Then, we construct an auxiliary function, called a discrete filled function, at this local minimizer. By minimizing the filled function, either an improved local minimizer is found or one of the vertices of the constraint set is reached. Otherwise, the parameters of the filled function are adjusted. This process is repeated until no better local minimizer of the corresponding filled function is found. The final local minimizer is then taken as an approximation of the global minimizer.While the main aim of this thesis is to present a new computational methodfor solving discrete-valued optimal control problems, the initial focus is on solvingpurely discrete optimization problems. We identify several discrete filled functionstechniques in the literature and perform a critical review including comprehensive numerical tests. Once the best filled function method is identified, we propose and test several variations of the method with numerical examples.We then consider the task of determining near globally optimal solutions of discrete-valued optimal control problems. The main difficulty in solving the discrete-valued optimal control problems is that the control restraint set is discrete and hence not convex. Conventional computational optimal control techniques are designed for problems in which the control takes values in a connected set, such as an interval, and thus they cannot solve the problem directly. Furthermore, variable switching times are known to cause problems in the implementation of any numerical algorithm due to the variable location of discontinuities in the dynamics. Therefore, such problem cannot be solved using conventional computational approaches. We propose a time scaling transformation to overcome this difficulty, where a new discrete variable representing the switching sequence and a new variable controlling the switching times are introduced. The transformation results in an equivalent mixed discrete optimization problem. The transformed problemis then decomposed into a bi-level optimization problem, which is solved using a combination of an efficient discrete filled function method identified earlier and a computational optimal control technique based on the concept of control parameterization.To demonstrate the applicability of the proposed method, we solve two complex applied engineering problems involving a hybrid power system and a sensor scheduling task, respectively. Computational results indicate that this method is robust, reliable, and efficient. It can successfully identify a near-global solution for these complex applied optimization problems, despite the demonstrated presence of multiple local optima. In addition, we also compare the results obtained with other methods in the literature. Numerical results confirm that the proposed method yields significant improvements over those obtained by other methods

    Differential Evolution Methods for the Fuzzy Extension of Functions

    Get PDF
    The paper illustrates a differential evolution (DE) algorithm to calculate the level-cuts of the fuzzy extension of a multidimensional real valued function to fuzzy numbers. The method decomposes the fuzzy extension engine into a set of "nested" min and max box-constrained op- timization problems and uses a form of the DE algorithm, based on multi populations which cooperate during the search phase and specialize, a part of the populations to find the the global min (corresponding to lower branch of the fuzzy extension) and a part of the populations to find the global max (corresponding to the upper branch), both gaining efficiency from the work done for a level-cut to the subsequent ones. A special ver- sion of the algorithm is designed to the case of differentiable functions, for which a representation of the fuzzy numbers is used to improve ef- ficiency and quality of calculations. The included computational results indicate that the DE method is a promising tool as its computational complexity grows on average superlinearly (of degree less than 1.5) in the number of variables of the function to be extended.Fuzzy Sets, Differential Evolution Method, Fuzzy Extension of Functions
    • …
    corecore