2,760 research outputs found
A Genetic Programming Approach to Designing Convolutional Neural Network Architectures
The convolutional neural network (CNN), which is one of the deep learning
models, has seen much success in a variety of computer vision tasks. However,
designing CNN architectures still requires expert knowledge and a lot of trial
and error. In this paper, we attempt to automatically construct CNN
architectures for an image classification task based on Cartesian genetic
programming (CGP). In our method, we adopt highly functional modules, such as
convolutional blocks and tensor concatenation, as the node functions in CGP.
The CNN structure and connectivity represented by the CGP encoding method are
optimized to maximize the validation accuracy. To evaluate the proposed method,
we constructed a CNN architecture for the image classification task with the
CIFAR-10 dataset. The experimental result shows that the proposed method can be
used to automatically find the competitive CNN architecture compared with
state-of-the-art models.Comment: This is the revised version of the GECCO 2017 paper. The code of our
method is available at https://github.com/sg-nm/cgp-cn
Large-Batch, Neural Multi-Objective Bayesian Optimization
Bayesian optimization provides a powerful framework for global optimization
of black-box, expensive-to-evaluate functions. However, it has a limited
capacity in handling data-intensive problems, especially in multi-objective
settings, due to the poor scalability of default Gaussian Process surrogates.
We present a novel Bayesian optimization framework specifically tailored to
address these limitations. Our method leverages a Bayesian neural networks
approach for surrogate modeling. This enables efficient handling of large
batches of data, modeling complex problems, and generating the uncertainty of
the predictions. In addition, our method incorporates a scalable,
uncertainty-aware acquisition strategy based on the well-known, easy-to-deploy
NSGA-II. This fully parallelizable strategy promotes efficient exploration of
uncharted regions. Our framework allows for effective optimization in
data-intensive environments with a minimum number of iterations. We demonstrate
the superiority of our method by comparing it with state-of-the-art
multi-objective optimizations. We perform our evaluation on two real-world
problems - airfoil design and color printing - showcasing the applicability and
efficiency of our approach. Code is available at:
https://github.com/an-on-ym-ous/lbn\_mob
ϵ-shotgun: ϵ-greedy batch bayesian optimisation
Bayesian optimisation is a popular surrogate model-based approach for optimising expensive black-box functions. Given a surrogate model, the next location to expensively evaluate is chosen via maximisation of a cheap-to-query acquisition function. We present an ϵ-greedy procedure for Bayesian optimisation in batch settings in which the black-box function can be evaluated multiple times in parallel. Our ϵ-shotgun algorithm leverages the model's prediction, uncertainty, and the approximated rate of change of the landscape to determine the spread of batch solutions to be distributed around a putative location. The initial target location is selected either in an exploitative fashion on the mean prediction, or - with probability ϵ - from elsewhere in the design space. This results in locations that are more densely sampled in regions where the function is changing rapidly and in locations predicted to be good (i.e. close to predicted optima), with more scattered samples in regions where the function is flatter and/or of poorer quality. We empirically evaluate the ϵ-shotgun methods on a range of synthetic functions and two real-world problems, finding that they perform at least as well as state-of-the-art batch methods and in many cases exceed their performance
A portfolio approach to massively parallel Bayesian optimization
One way to reduce the time of conducting optimization studies is to evaluate
designs in parallel rather than just one-at-a-time. For expensive-to-evaluate
black-boxes, batch versions of Bayesian optimization have been proposed. They
work by building a surrogate model of the black-box that can be used to select
the designs to evaluate efficiently via an infill criterion. Still, with higher
levels of parallelization becoming available, the strategies that work for a
few tens of parallel evaluations become limiting, in particular due to the
complexity of selecting more evaluations. It is even more crucial when the
black-box is noisy, necessitating more evaluations as well as repeating
experiments. Here we propose a scalable strategy that can keep up with massive
batching natively, focused on the exploration/exploitation trade-off and a
portfolio allocation. We compare the approach with related methods on
deterministic and noisy functions, for mono and multiobjective optimization
tasks. These experiments show similar or better performance than existing
methods, while being orders of magnitude faster
- …