204,581 research outputs found

    Active Bayesian Optimization: Minimizing Minimizer Entropy

    Full text link
    The ultimate goal of optimization is to find the minimizer of a target function.However, typical criteria for active optimization often ignore the uncertainty about the minimizer. We propose a novel criterion for global optimization and an associated sequential active learning strategy using Gaussian processes.Our criterion is the reduction of uncertainty in the posterior distribution of the function minimizer. It can also flexibly incorporate multiple global minimizers. We implement a tractable approximation of the criterion and demonstrate that it obtains the global minimizer accurately compared to conventional Bayesian optimization criteria

    Efficient Graph-Based Active Learning with Probit Likelihood via Gaussian Approximations

    Full text link
    We present a novel adaptation of active learning to graph-based semi-supervised learning (SSL) under non-Gaussian Bayesian models. We present an approximation of non-Gaussian distributions to adapt previously Gaussian-based acquisition functions to these more general cases. We develop an efficient rank-one update for applying "look-ahead" based methods as well as model retraining. We also introduce a novel "model change" acquisition function based on these approximations that further expands the available collection of active learning acquisition functions for such methods.Comment: Accepted in ICML Workshop on Real World Experiment Design and Active Learning 202

    Robustness of computer algorithms to simulate optimal experimentation problems.

    Get PDF
    Three methods have been developed by the authors for solving optimal experimentation problems. David Kendrick (1981, 2002, Ch.10) uses quadratic approximation of the value function and linear approximation of the equation of motion to simulate general optimal experimentation (active learning) problems. Beck and Volker Wieland (2002) use dynamic programming methods to develop an algorithm for optimal experimentation problems. Cosimano (2003) and Cosimano and Gapen (2005) use the Perturbation method to develop an algorithm for solving optimal experimentation problems. The perturbation is in the neighborhood of the augmented linear regulator problems of Hansen and Sargent (2004). In this paper we take an example from Beck and Wieland which fits into the setup of all three algorithms. Using this example we examine the cost and benefits of the various algorithms for solving optimal experimentation problems.

    Bayesian active learning line sampling with log-normal process for rare-event probability estimation

    Get PDF
    Line sampling (LS) stands as a powerful stochastic simulation method for structural reliability analysis, especially for assessing small failure probabilities. To further improve the performance of traditional LS, a Bayesian active learning idea has recently been pursued. This work presents another Bayesian active learning alternative, called ‘Bayesian active learning line sampling with log-normal process’ (BAL-LS-LP), to traditional LS. In this method, we assign an LP prior instead of a Gaussian process prior over the distance function so as to account for its non-negativity constraint. Besides, the approximation error between the logarithmic approximate distance function and the logarithmic true distance function is assumed to follow a zero-mean normal distribution. The approximate posterior mean and variance of the failure probability are derived accordingly. Based on the posterior statistics of the failure probability, a learning function and a stopping criterion are developed to enable Bayesian active learning. In the numerical implementation of the proposed BAL-LS-LP method, the important direction can be updated on the fly without re-evaluating the distance function. Four numerical examples are studied to demonstrate the proposed method. Numerical results show that the proposed method can estimate extremely small failure probabilities with desired efficiency and accuracy

    Submodularity in Batch Active Learning and Survey Problems on Gaussian Random Fields

    Full text link
    Many real-world datasets can be represented in the form of a graph whose edge weights designate similarities between instances. A discrete Gaussian random field (GRF) model is a finite-dimensional Gaussian process (GP) whose prior covariance is the inverse of a graph Laplacian. Minimizing the trace of the predictive covariance Sigma (V-optimality) on GRFs has proven successful in batch active learning classification problems with budget constraints. However, its worst-case bound has been missing. We show that the V-optimality on GRFs as a function of the batch query set is submodular and hence its greedy selection algorithm guarantees an (1-1/e) approximation ratio. Moreover, GRF models have the absence-of-suppressor (AofS) condition. For active survey problems, we propose a similar survey criterion which minimizes 1'(Sigma)1. In practice, V-optimality criterion performs better than GPs with mutual information gain criteria and allows nonuniform costs for different nodes

    Non-monotone Submodular Maximization with Nearly Optimal Adaptivity and Query Complexity

    Full text link
    Submodular maximization is a general optimization problem with a wide range of applications in machine learning (e.g., active learning, clustering, and feature selection). In large-scale optimization, the parallel running time of an algorithm is governed by its adaptivity, which measures the number of sequential rounds needed if the algorithm can execute polynomially-many independent oracle queries in parallel. While low adaptivity is ideal, it is not sufficient for an algorithm to be efficient in practice---there are many applications of distributed submodular optimization where the number of function evaluations becomes prohibitively expensive. Motivated by these applications, we study the adaptivity and query complexity of submodular maximization. In this paper, we give the first constant-factor approximation algorithm for maximizing a non-monotone submodular function subject to a cardinality constraint kk that runs in O(log(n))O(\log(n)) adaptive rounds and makes O(nlog(k))O(n \log(k)) oracle queries in expectation. In our empirical study, we use three real-world applications to compare our algorithm with several benchmarks for non-monotone submodular maximization. The results demonstrate that our algorithm finds competitive solutions using significantly fewer rounds and queries.Comment: 12 pages, 8 figure
    corecore