64,194 research outputs found
Domain-Agnostic Batch Bayesian Optimization with Diverse Constraints via Bayesian Quadrature
Real-world optimisation problems often feature complex combinations of (1)
diverse constraints, (2) discrete and mixed spaces, and are (3) highly
parallelisable. (4) There are also cases where the objective function cannot be
queried if unknown constraints are not satisfied, e.g. in drug discovery,
safety on animal experiments (unknown constraints) must be established before
human clinical trials (querying objective function) may proceed. However, most
existing works target each of the above three problems in isolation and do not
consider (4) unknown constraints with query rejection. For problems with
diverse constraints and/or unconventional input spaces, it is difficult to
apply these techniques as they are often mutually incompatible. We propose
cSOBER, a domain-agnostic prudent parallel active sampler for Bayesian
optimisation, based on SOBER of Adachi et al. (2023). We consider infeasibility
under unknown constraints as a type of integration error that we can estimate.
We propose a theoretically-driven approach that propagates such error as a
tolerance in the quadrature precision that automatically balances exploitation
and exploration with the expected rejection rate. Moreover, our method flexibly
accommodates diverse constraints and/or discrete and mixed spaces via adaptive
tolerance, including conventional zero-risk cases. We show that cSOBER
outperforms competitive baselines on diverse real-world blackbox-constrained
problems, including safety-constrained drug discovery, and
human-relationship-aware team optimisation over graph-structured space.Comment: 24 pages, 5 figure
Offline Learning for Sequence-based Selection Hyper-heuristics
This thesis is concerned with finding solutions to discrete NP-hard problems. Such problems occur in a wide range of real-world applications, such as bin packing, industrial flow shop problems, determining Boolean satisfiability, the traveling salesman and vehicle routing problems, course timetabling, personnel scheduling, and the optimisation of water distribution networks. They are typically represented as optimisation problems where the goal is to find a ``best'' solution from a given space of feasible solutions. As no known polynomial-time algorithmic solution exists for NP-hard problems, they are usually solved by applying heuristic methods. Selection hyper-heuristics are algorithms that organise and combine a number of individual low level heuristics into a higher level framework with the objective of improving optimisation performance. Many selection hyper-heuristics employ learning algorithms in order to enhance optimisation performance by improving the selection of single heuristics, and this learning may be classified as either online or offline. This thesis presents a novel statistical framework for the offline learning of subsequences of low level heuristics in order to improve the optimisation performance of sequenced-based selection hyper-heuristics. A selection hyper-heuristic is used to optimise the HyFlex set of discrete benchmark problems. The resulting sequences of low level heuristic selections and objective function values are used to generate an offline learning database of heuristic selections. The sequences in the database are broken down into subsequences and the mathematical concept of a logarithmic return is used to discriminate between ``effective'' subsequences, that tend to lead to improvements in optimisation performance, and ``disruptive'' subsequences that tend to lead to worsening performance. Effective subsequences are used to improve hyper-heuristics performance directly, by embedding them in a simple hyper-heuristic design, and indirectly as the inputs to an appropriate hyper-heuristic learning algorithm. Furthermore, by comparing effective subsequences across different problem domains it is possible to investigate the potential for cross-domain learning. The results presented here demonstrates that the use of well chosen subsequences of heuristics can lead to small, but statistically significant, improvements in optimisation performance
Hybrid evolutionary techniques for constrained optimisation design
This thesis a research program in which novel and generic optimisation methods were developed so that can be applied to a multitude of mathematically modelled business problems which the standard optimisation techniques often fail to deal with. The continuous and mixed discrete optimisation methods have been investigated by designing new approaches that allow users to more effectively tackle difficult optimisation problems with a mix of integer and real valued variables. The focus of this thesis presents practical suggestions towards the implementation of hybrid evolutionary approaches for solving optimisation problems with highly structured constraints. This work also introduces a derivation of the different optimisation methods that have been reported in the literature. Major theoretical properties of the new methods have been presented and implemented. Here we present detailed description of the most essential steps of the implementation. The performance of the developed methods is evaluated against real-world benchmark problems, and the numerical results of the test problems are found to be competitive compared to existing methods
Modelling and solution methods for portfolio optimisation
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University, 16/01/2004.In this thesis modelling and solution methods for portfolio optimisation are presented. The investigations reported in this thesis extend the Markowitz mean-variance model to the domain of quadratic mixed integer programming (QMIP) models which are 'NP-hard' discrete optimisation problems. In addition to the modelling extensions a number of challenging aspects of solution algorithms are considered. The relative performances of sparse simplex (SSX) as well as the interior point method (IPM) are studied in detail. In particular, the roles of 'warmstart' and dual simplex are highlighted as applied to the construction of the efficient frontier which requires processing a family of problems; that is, the portfolio planning model stated in a parametric form. The method of solving QMIP models using the branch and bound algorithm is first developed; this is followed up by heuristics which improve the performance of the (discrete) solution algorithm. Some properties of the efficient frontier with discrete constraints are considered and a method of computing the discrete efficient frontier (DEF) efficiently is proposed. The computational investigation considers the efficiency and effectiveness in respect of the scale up properties of the proposed algorithm. The extensions of the real world models and the proposed solution algorithms make contribution as new knowledge
On discretisation drift and smoothness regularisation in neural network training
The deep learning recipe of casting real-world problems as mathematical
optimisation and tackling the optimisation by training deep neural networks
using gradient-based optimisation has undoubtedly proven to be a fruitful one.
The understanding behind why deep learning works, however, has lagged behind
its practical significance. We aim to make steps towards an improved
understanding of deep learning with a focus on optimisation and model
regularisation. We start by investigating gradient descent (GD), a
discrete-time algorithm at the basis of most popular deep learning optimisation
algorithms. Understanding the dynamics of GD has been hindered by the presence
of discretisation drift, the numerical integration error between GD and its
often studied continuous-time counterpart, the negative gradient flow (NGF). To
add to the toolkit available to study GD, we derive novel continuous-time flows
that account for discretisation drift. Unlike the NGF, these new flows can be
used to describe learning rate specific behaviours of GD, such as training
instabilities observed in supervised learning and two-player games. We then
translate insights from continuous time into mitigation strategies for unstable
GD dynamics, by constructing novel learning rate schedules and regularisers
that do not require additional hyperparameters. Like optimisation, smoothness
regularisation is another pillar of deep learning's success with wide use in
supervised learning and generative modelling. Despite their individual
significance, the interactions between smoothness regularisation and
optimisation have yet to be explored. We find that smoothness regularisation
affects optimisation across multiple deep learning domains, and that
incorporating smoothness regularisation in reinforcement learning leads to a
performance boost that can be recovered using adaptions to optimisation
methods.Comment: PhD thesis. arXiv admin note: text overlap with arXiv:2302.0195
Recommended from our members
On the effect of fluid-structure interactions and choice of algorithm in multi-physics topology optimisation
This article presents an optimisation framework for the compliance minimisation of structures subjected to design-dependent pressure loads. A finite element solver coupled to a Lattice Boltzmann method is employed, such that the effect of the fluid-structure interactions on the optimised design can be considered. It is noted that the main computational expense of the algorithm
is the Lattice Boltzmann method. Therefore, to improve the computational
efficiency and to assess the effect of the fluid-structure interactions on the fi nal optimised design, the degree of coupling is changed.
Several successful topology optimisation algorithms exist with thousands
of associated publications in the literature. However, only a small portion of these are applied to real-world problems, with even fewer offering a comparison of methodologies. This is especially important for problems involving fluid-structure interactions, where discrete and continuous methods can provide different advantages.
The goal of this research is to couple two key disciplines, fluids and structures, into a topology optimisation framework, which shows fast convergence for multi-physics optimisation problems. This is achieved by offering a comparison of three popular, but competing, optimisation methodologies. The needs for the exploration of larger design spaces and to produce innovative designs make meta-heuristic algorithms less efficient for this task. A coupled analysis, where the fluid and structural mechanics are updated, provides superior results compared with an uncoupled analysis approach, however at some computational expense. The results in this article show that the method is sensitive to whether fluid-structure coupling is included, i.e. if the fluid mechanics are updated with design changes, but not to the degree of the coupling, i.e. how regularly the fluid mechanics are updated, up to a certain limit. Therefore, the computational efficiency of the algorithm can be considerably increased with small penalties in the quality of the objective by relaxing the coupling
On discretisation drift and smoothness regularisation in neural network training
The deep learning recipe of casting real-world problems as mathematical optimisation and tackling the optimisation by training deep neural networks using gradient-based optimisation has undoubtedly proven to be a fruitful one. The understanding behind why deep learning works, however, has lagged behind its practical significance. We aim to make steps towards an improved understanding of deep learning with a focus on optimisation and model regularisation. We start by investigating gradient descent (GD), a discrete-time algorithm at the basis of most popular deep learning optimisation algorithms. Understanding the dynamics of GD has been hindered by the presence of discretisation drift, the numerical integration error between GD and its often studied continuous-time counterpart, the negative gradient flow (NGF). To add to the toolkit available to study GD, we derive novel continuous-time flows that account for discretisation drift. Unlike the NGF, these new flows can be used to describe learning rate specific behaviours of GD, such as training instabilities observed in supervised learning and two-player games. We then translate insights from continuous time into mitigation strategies for unstable GD dynamics, by constructing novel learning rate schedules and regularisers that do not require additional hyperparameters. Like optimisation, smoothness regularisation is another pillar of deep learning's success with wide use in supervised learning and generative modelling. Despite their individual significance, the interactions between smoothness regularisation and optimisation have yet to be explored. We find that smoothness regularisation affects optimisation across multiple deep learning domains, and that incorporating smoothness regularisation in reinforcement learning leads to a performance boost that can be recovered using adaptions to optimisation methods
Recommended from our members
New variants of variable neighbourhood search for 0-1 mixed integer programming and clustering
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.Many real-world optimisation problems are discrete in nature. Although recent rapid developments in computer technologies are steadily increasing the speed of computations, the size of an instance of a hard discrete optimisation problem solvable in prescribed time does not increase linearly with the computer speed. This calls for the development of new solution methodologies for solving larger instances in shorter time. Furthermore, large instances of discrete optimisation problems are normally impossible to solve to optimality within a reasonable computational time/space and can only be tackled with a heuristic approach.
In this thesis the development of so called matheuristics, the heuristics which are based on the mathematical formulation of the problem, is studied and employed within the variable neighbourhood search framework. Some new variants of the variable neighbourhood searchmetaheuristic itself are suggested, which naturally emerge from exploiting the information from the mathematical programming formulation of the problem. However, those variants may also be applied to problems described by the combinatorial formulation. A unifying perspective on modern advances in local search-based metaheuristics, a so called hyper-reactive approach, is also proposed. Two NP-hard discrete optimisation problems are considered: 0-1 mixed integer programming and clustering with application to colour image quantisation. Several new heuristics for 0-1 mixed integer programming problem are developed, based on the principle of variable neighbourhood search. One set of proposed heuristics consists of improvement heuristics, which attempt to find high-quality near-optimal solutions starting from a given feasible solution. Another set consists of constructive heuristics, which attempt to find initial feasible solutions for 0-1 mixed integer programs. Finally, some variable neighbourhood search based clustering techniques are applied for solving the colour image quantisation problem. All new methods presented are compared to other algorithms recommended in literature and a comprehensive performance analysis is provided. Computational results show that the methods proposed either outperform the existing state-of-the-art methods for the problems observed, or provide comparable results.
The theory and algorithms presented in this thesis indicate that hybridisation of the CPLEX MIP solver and the VNS metaheuristic can be very effective for solving large instances of the 0-1 mixed integer programming problem. More generally, the results presented in this thesis suggest that hybridisation of exact (commercial) integer programming solvers and some metaheuristic methods is of high interest and such combinations deserve further practical and theoretical investigation. Results also show that VNS can be successfully applied to solving a colour image quantisation problem.Support from the Mathematical Institute, Serbian Academy of Sciences and Arts, are acknowledged for this research
- …