264 research outputs found
Random gradient-free minimization of convex functions
In this paper, we prove the complexity bounds for methods of Convex Optimization based only on computation of the function value. The search directions of our schemes are normally distributed random Gaussian vectors. It appears that such methods usually need at most n times more iterations than the standard gradient methods, where n is the dimension of the space of variables. This conclusion is true both for nonsmooth and smooth problems. For the later class, we present also an accelerated scheme with the expected rate of convergence O(n[ exp ]2 /k[ exp ]2), where k is the iteration counter. For Stochastic Optimization, we propose a zero-order scheme and justify its expected rate of convergence O(n/k[ exp ]1/2). We give also some bounds for the rate of convergence of the random gradient-free methods to stationary points of nonconvex functions, both for smooth and nonsmooth cases. Our theoretical results are supported by preliminary computational experiments.convex optimization, stochastic optimization, derivative-free methods, random methods, complexity bounds
Exploiting higher order smoothness in derivative-free optimization and continuous bandits
We study the problem of zero-order optimization of a strongly convex function. The goal is to find the minimizer of the function by a sequential exploration of its values, under measurement noise. We study the impact of higher order smoothness properties of the function on the optimization error and on the cumulative regret. To solve this problem we consider a randomized approximation of the projected gradient descent algorithm. The gradient is estimated by a randomized procedure involving two function evaluations and a smoothing kernel. We derive upper bounds for this algorithm both in the constrained and unconstrained settings and prove minimax lower bounds for any sequential search method. Our results imply that the zero-order algorithm is nearly optimal in terms of sample complexity and the problem parameters. Based on this algorithm, we also propose an estimator of the minimum value of the function achieving almost sharp oracle behavior. We compare our results with the state-of-the-art, highlighting a number of key improvements
PyPop7: A Pure-Python Library for Population-Based Black-Box Optimization
In this paper, we present a pure-Python open-source library, called PyPop7,
for black-box optimization (BBO). It provides a unified and modular interface
for more than 60 versions and variants of different black-box optimization
algorithms, particularly population-based optimizers, which can be classified
into 12 popular families: Evolution Strategies (ES), Natural Evolution
Strategies (NES), Estimation of Distribution Algorithms (EDA), Cross-Entropy
Method (CEM), Differential Evolution (DE), Particle Swarm Optimizer (PSO),
Cooperative Coevolution (CC), Simulated Annealing (SA), Genetic Algorithms
(GA), Evolutionary Programming (EP), Pattern Search (PS), and Random Search
(RS). It also provides many examples, interesting tutorials, and full-fledged
API documentations. Through this new library, we expect to provide a
well-designed platform for benchmarking of optimizers and promote their
real-world applications, especially for large-scale BBO. Its source code and
documentations are available at
https://github.com/Evolutionary-Intelligence/pypop and
https://pypop.readthedocs.io/en/latest, respectively.Comment: 5 page
- …