204,581 research outputs found
Active Bayesian Optimization: Minimizing Minimizer Entropy
The ultimate goal of optimization is to find the minimizer of a target
function.However, typical criteria for active optimization often ignore the
uncertainty about the minimizer. We propose a novel criterion for global
optimization and an associated sequential active learning strategy using
Gaussian processes.Our criterion is the reduction of uncertainty in the
posterior distribution of the function minimizer. It can also flexibly
incorporate multiple global minimizers. We implement a tractable approximation
of the criterion and demonstrate that it obtains the global minimizer
accurately compared to conventional Bayesian optimization criteria
Efficient Graph-Based Active Learning with Probit Likelihood via Gaussian Approximations
We present a novel adaptation of active learning to graph-based
semi-supervised learning (SSL) under non-Gaussian Bayesian models. We present
an approximation of non-Gaussian distributions to adapt previously
Gaussian-based acquisition functions to these more general cases. We develop an
efficient rank-one update for applying "look-ahead" based methods as well as
model retraining. We also introduce a novel "model change" acquisition function
based on these approximations that further expands the available collection of
active learning acquisition functions for such methods.Comment: Accepted in ICML Workshop on Real World Experiment Design and Active
Learning 202
Robustness of computer algorithms to simulate optimal experimentation problems.
Three methods have been developed by the authors for solving optimal experimentation problems. David Kendrick (1981, 2002, Ch.10) uses quadratic approximation of the value function and linear approximation of the equation of motion to simulate general optimal experimentation (active learning) problems. Beck and Volker Wieland (2002) use dynamic programming methods to develop an algorithm for optimal experimentation problems. Cosimano (2003) and Cosimano and Gapen (2005) use the Perturbation method to develop an algorithm for solving optimal experimentation problems. The perturbation is in the neighborhood of the augmented linear regulator problems of Hansen and Sargent (2004). In this paper we take an example from Beck and Wieland which fits into the setup of all three algorithms. Using this example we examine the cost and benefits of the various algorithms for solving optimal experimentation problems.
Bayesian active learning line sampling with log-normal process for rare-event probability estimation
Line sampling (LS) stands as a powerful stochastic simulation method for structural reliability analysis, especially for assessing small failure probabilities. To further improve the performance of traditional LS, a Bayesian active learning idea has recently been pursued. This work presents another Bayesian active learning alternative, called ‘Bayesian active learning line sampling with log-normal process’ (BAL-LS-LP), to traditional LS. In this method, we assign an LP prior instead of a Gaussian process prior over the distance function so as to account for its non-negativity constraint. Besides, the approximation error between the logarithmic approximate distance function and the logarithmic true distance function is assumed to follow a zero-mean normal distribution. The approximate posterior mean and variance of the failure probability are derived accordingly. Based on the posterior statistics of the failure probability, a learning function and a stopping criterion are developed to enable Bayesian active learning. In the numerical implementation of the proposed BAL-LS-LP method, the important direction can be updated on the fly without re-evaluating the distance function. Four numerical examples are studied to demonstrate the proposed method. Numerical results show that the proposed method can estimate extremely small failure probabilities with desired efficiency and accuracy
Submodularity in Batch Active Learning and Survey Problems on Gaussian Random Fields
Many real-world datasets can be represented in the form of a graph whose edge
weights designate similarities between instances. A discrete Gaussian random
field (GRF) model is a finite-dimensional Gaussian process (GP) whose prior
covariance is the inverse of a graph Laplacian. Minimizing the trace of the
predictive covariance Sigma (V-optimality) on GRFs has proven successful in
batch active learning classification problems with budget constraints. However,
its worst-case bound has been missing. We show that the V-optimality on GRFs as
a function of the batch query set is submodular and hence its greedy selection
algorithm guarantees an (1-1/e) approximation ratio. Moreover, GRF models have
the absence-of-suppressor (AofS) condition. For active survey problems, we
propose a similar survey criterion which minimizes 1'(Sigma)1. In practice,
V-optimality criterion performs better than GPs with mutual information gain
criteria and allows nonuniform costs for different nodes
Non-monotone Submodular Maximization with Nearly Optimal Adaptivity and Query Complexity
Submodular maximization is a general optimization problem with a wide range
of applications in machine learning (e.g., active learning, clustering, and
feature selection). In large-scale optimization, the parallel running time of
an algorithm is governed by its adaptivity, which measures the number of
sequential rounds needed if the algorithm can execute polynomially-many
independent oracle queries in parallel. While low adaptivity is ideal, it is
not sufficient for an algorithm to be efficient in practice---there are many
applications of distributed submodular optimization where the number of
function evaluations becomes prohibitively expensive. Motivated by these
applications, we study the adaptivity and query complexity of submodular
maximization. In this paper, we give the first constant-factor approximation
algorithm for maximizing a non-monotone submodular function subject to a
cardinality constraint that runs in adaptive rounds and makes
oracle queries in expectation. In our empirical study, we use
three real-world applications to compare our algorithm with several benchmarks
for non-monotone submodular maximization. The results demonstrate that our
algorithm finds competitive solutions using significantly fewer rounds and
queries.Comment: 12 pages, 8 figure
- …