7 research outputs found

    A derivative-free filter driven multistart technique for global optimization

    Get PDF
    A stochastic global optimization method based on a multistart strategy and a derivative-free filter local search for general constrained optimization is presented and analyzed. In the local search procedure, approximate descent directions for the constraint violation or the objective function are used to progress towards the optimal solution. The algorithm is able to locate all the local minima, and consequently, the global minimum of a multi-modal objective function. The performance of the multistart method is analyzed with a set of benchmark problems and a comparison is made with other methods.This work was financed by FEDER funds through COMPETE-Programa Operacional Fatores de Competitividade and by portuguese funds through FCT-Fundação para a Ciência e a Tecnologia within projects PEst-C/MAT/UI0013/2011 and FCOMP- 01-0124-FEDER-022674

    Improving efficiency of a multistart with interrupted Hooke-and-Jeeves filter search for solving MINLP problems

    Get PDF
    Publicado em: "Computational science and its applications – ICCSA 2016: 16th International Conference, Beijing, China, July 4-7, 2016, Proceedings, Part I". ISBN 978-3-319-42084-4This paper addresses the problem of solving mixed-integer nonlinear programming (MINLP) problems by a multistart strategy that invokes a derivative-free local search procedure based on a filter set methodology to handle nonlinear constraints. A new concept of componentwise normalized distance aiming to discard randomly generated points that are sufficiently close to other points already used to invoke the local search is analyzed. A variant of the Hooke-and-Jeeves filter algorithm for MINLP is proposed with the goal of interrupting the iterative process if the accepted iterate falls inside an -neighborhood of an already computed minimizer. Preliminary numerical results are included.FCT - Fundação para a Ciência e Tecnologia, within the projects UID/CEC/00319/2013 and UID/MAT/00013/2013.COMPETE: POCI-01- 0145-FEDER-00704

    Multiple roots of systems of equations by repulsion merit functions

    Get PDF
    In this paper we address the problem of computing multiple roots of a system of nonlinear equations through the global optimization of an appropriate merit function. The search procedure for a global min- imizer of the merit function is carried out by a metaheuristic, known as harmony search, which does not require any derivative information. The multiple roots of the system are sequentially determined along several ite- rations of a single run, where the merit function is accordingly modified by penalty terms that aim to create repulsion areas around previously computed minimizers. A repulsion algorithm based on a multiplicative kind penalty function is proposed. Preliminary numerical experiments with a benchmark set of problems show the effectiveness of the proposed method.Fundação para a Ciência e a Tecnologia (FCT

    Multiple solutions of mixed variable optimization by multistart hooke and jeeves filter method

    Get PDF
    In this study, we propose a multistart method based on an extended version of the Hooke and Jeeves (HJ) algorithm for computing mul- tiple solutions of mixed variable optimization problems. The inequal- ity and equality constraints of the problem are handled by a filter set methodology. The basic ideas present in the HJ algorithm, namely the exploratory and pattern moves, are extended to consider two objective functions and to handle continuous and integer variables simultaneously. This proposal is integrated into a multistart method as a local search procedure that is repeatedly invoked to converge to different global and non-global optimal solutions starting from randomly generated points. To avoid repeated convergence to previously computed solutions, the concept of region of attraction of an optimizer is implemented. The performance of the new method is tested on benchmark problems. Its effectiveness is emphasized by a comparison with a well-known solver.Fundação para a Ciência e a Tecnologia (FCT

    Multilocal programming and applications

    Get PDF
    Preprint versionMultilocal programming aims to identify all local minimizers of unconstrained or constrained nonlinear optimization problems. The multilocal programming theory relies on global optimization strategies combined with simple ideas that are inspired in deflection or stretching techniques to avoid convergence to the already detected local minimizers. The most used methods to solve this type of problems are based on stochastic procedures and a population of solutions. In general, population-based methods are computationally expensive but rather reliable in identifying all local solutions. In this chapter, a review on recent techniques for multilocal programming is presented. Some real-world multilocal programming problems based on chemical engineering process design applications are described.Fundação para a Ciência e a Tecnologia (FCT

    Self-adaptive combination of global tabu search and local search for nonlinear equations

    Get PDF
    Solving systems of nonlinear equations is a very important task since the problems emerge mostly through the mathematical modeling of real problems that arise naturally in many branches of engineering and in the physical sciences. The problem can be naturally reformulated as a global optimization problem. In this paper, we show that a self-adaptive combination of a metaheuristic with a classical local search method is able to converge to some difficult problems that are not solved by Newton-type methodsFundação para a Ciência e a Tecnologia (FCT

    Bounding the search space for global optimization of neural networks learning error: an interval analysis approach

    Get PDF
    Training a multilayer perceptron (MLP) with algorithms employing global search strategies has been an important research direction in the field of neural networks. Despite a number of significant results, an important matter concerning the bounds of the search region---typically defined as a box---where a global optimization method has to search for a potential global minimizer seems to be unresolved. The approach presented in this paper builds on interval analysis and attempts to define guaranteed bounds in the search space prior to applying a global search algorithm for training an MLP. These bounds depend on the machine precision and the term guaranteed denotes that the region defined surely encloses weight sets that are global minimizers of the neural network's error function. Although the solution set to the bounding problem for an MLP is in general non-convex, the paper presents the theoretical results that help deriving a box which is a convex set. This box is an outer approximation of the algebraic solutions to the interval equations resulting from the function implemented by the network nodes. An experimental study using well known benchmarks is presented in accordance with the theoretical results
    corecore