64 research outputs found
The Krigifier: A Procedure for Generating Pseudorandom Nonlinear Objective Functions for Computational Experimentation
Comprehensive computational experiments to assess the performance of algorithms for numerical optimization require (among other things) a practical procedure for generating pseudorandom nonlinear objective functions. We propose a procedure that is based on the convenient fiction that objective functions are realizations of stochastic processes. This report details the calculations necessary to implement our procedure for the case of certain stationary Gaussian processes and presents a specific implementation in the statistical programming language S-PLUS
Parallel Deterministic and Stochastic Global Minimization of Functions with Very Many Minima
The optimization of three problems with high dimensionality and many local minima are investigated
under five different optimization algorithms: DIRECT, simulated annealing, Spall’s SPSA algorithm, the KNITRO
package, and QNSTOP, a new algorithm developed at Indiana University
Adjusting process count on demand for petascale global optimization⋆
There are many challenges that need to be met before efficient and reliable computation at the
petascale is possible. Many scientific and engineering codes running at the petascale are likely to
be memory intensive, which makes thrashing a serious problem for many petascale applications.
One way to overcome this challenge is to use a dynamic number of processes, so that the total
amount of memory available for the computation can be increased on demand. This paper
describes modifications made to the massively parallel global optimization code pVTdirect in
order to allow for a dynamic number of processes. In particular, the modified version of the
code monitors memory use and spawns new processes if the amount of available memory is
determined to be insufficient. The primary design challenges are discussed, and performance
results are presented and analyzed
The Cost of Numerical Integration in Statistical Decision-theoretic Methods for Robust Design Optimization
The Bayes principle from statistical decision theory provides a conceptual framework
for quantifying uncertainties that arise in robust design optimization. The difficulty with
exploiting this framework is computational, as it leads to objective and constraint functions that
must be evaluated by numerical integration. Using a prototypical robust design optimization
problem, this study explores the computational cost of multidimensional integration (computing
expectation) and its interplay with optimization algorithms. It concludes that straightforward
application of standard off-the-shelf optimization software to robust design is prohibitively
expensive, necessitating adaptive strategies and the use of surrogates
- …