Optimal Anytime Search for Constrained Nonlinear Programming


In this thesis, we study optimal anytime stochastic search algorithms (SSAs) for solving general constrained nonlinear programming problems (NLPs) in discrete, continuous and mixed-integer space. The algorithms are general in the sense that they do not assume differentiability or convexity of functions. Based on the search algorithms, we develop the theory of SSAs and propose optimal SSAs with iterative deepening in order to minimize their expected search time. Based on the optimal SSAs, we then develop optimal anytime SSAs that generate improved solutions as more search time is allowed. Our SSAs for solving general constrained NLPs are based on the theory of discrete con-strained optimization using Lagrange multipliers that shows the equivalence between the set of constrained local minima (CLMdn) and the set of discrete-neighborhood saddle points (SPdn). To implement this theory, we propose a general procedural framework for locating an SPdn. By incorporating genetic algorithms in the framework, we evaluate new constrained search algorithms: constrained genetic algorithm (CGA) and combined constrained simulated annealing and genetic algorithm (CSAGA)

Similar works

Full text

oaioai:CiteSeerX.psu:10.1...Last time updated on 10/22/2014

This paper was published in CiteSeerX.

Having an issue?

Is data on this page outdated, violates copyrights or anything else? Report the problem now and we will take corresponding actions after reviewing your request.