6 research outputs found

    Reducing the Variance of Gaussian Process Hyperparameter Optimization with Preconditioning

    Full text link
    Gaussian processes remain popular as a flexible and expressive model class, but the computational cost of kernel hyperparameter optimization stands as a major limiting factor to their scaling and broader adoption. Recent work has made great strides combining stochastic estimation with iterative numerical techniques, essentially boiling down GP inference to the cost of (many) matrix-vector multiplies. Preconditioning -- a highly effective step for any iterative method involving matrix-vector multiplication -- can be used to accelerate convergence and thus reduce bias in hyperparameter optimization. Here, we prove that preconditioning has an additional benefit that has been previously unexplored. It not only reduces the bias of the log\log-marginal likelihood estimator and its derivatives, but it also simultaneously can reduce variance at essentially negligible cost. We leverage this result to derive sample-efficient algorithms for GP hyperparameter optimization requiring as few as O(log(ε1))\mathcal{O}(\log(\varepsilon^{-1})) instead of O(ε2)\mathcal{O}(\varepsilon^{-2}) samples to achieve error ε\varepsilon. Our theoretical results enable provably efficient and scalable optimization of kernel hyperparameters, which we validate empirically on a set of large-scale benchmark problems. There, variance reduction via preconditioning results in an order of magnitude speedup in hyperparameter optimization of exact GPs

    Finishing the euchromatic sequence of the human genome

    Get PDF
    The sequence of the human genome encodes the genetic instructions for human physiology, as well as rich information about human evolution. In 2001, the International Human Genome Sequencing Consortium reported a draft sequence of the euchromatic portion of the human genome. Since then, the international collaboration has worked to convert this draft into a genome sequence with high accuracy and nearly complete coverage. Here, we report the result of this finishing process. The current genome sequence (Build 35) contains 2.85 billion nucleotides interrupted by only 341 gaps. It covers ∼99% of the euchromatic genome and is accurate to an error rate of ∼1 event per 100,000 bases. Many of the remaining euchromatic gaps are associated with segmental duplications and will require focused work with new methods. The near-complete sequence, the first for a vertebrate, greatly improves the precision of biological analyses of the human genome including studies of gene number, birth and death. Notably, the human enome seems to encode only 20,000-25,000 protein-coding genes. The genome sequence reported here should serve as a firm foundation for biomedical research in the decades ahead

    Posterior and Computational Uncertainty in Gaussian Processes

    Full text link
    Gaussian processes scale prohibitively with the size of the dataset. In response, many approximation methods have been developed, which inevitably introduce approximation error. This additional source of uncertainty, due to limited computation, is entirely ignored when using the approximate posterior. Therefore in practice, GP models are often as much about the approximation method as they are about the data. Here, we develop a new class of methods that provides consistent estimation of the combined uncertainty arising from both the finite number of data observed and the finite amount of computation expended. The most common GP approximations map to an instance in this class, such as methods based on the Cholesky factorization, conjugate gradients, and inducing points. For any method in this class, we prove (i) convergence of its posterior mean in the associated RKHS, (ii) decomposability of its combined posterior covariance into mathematical and computational covariances, and (iii) that the combined variance is a tight worst-case bound for the squared error between the method's posterior mean and the latent function. Finally, we empirically demonstrate the consequences of ignoring computational uncertainty and show how implicitly modeling it improves generalization performance on benchmark datasets

    Multiple tracheal resections and anastomoses in a blue and gold macaw (Ara ararauna)

    No full text
    A 1.5-year-old, male blue and gold macaw (Ara ararauna) was anesthetized for a health examination and blood collection. The following day it was returned for episodes of coughing. The bird was presented again 13 days after the initial presentation with severe dyspnea. A tracheal stenosis was diagnosed by endoscopy and treated by surgical resection of 5 tracheal rings and tracheal anastomosis. The bird was discharged but returned 2 days later with a recurrent stenosis. Bougienage and balloon dilation of the stenotic area were performed separately; each resulted in less than 48 hours\u27 improvement in clinical signs after treatment. A second tracheal resection and anastomosis was performed, during which an additional 10 tracheal rings were removed. This second anastomosis was significantly more difficult to complete given the marked variation in diameter of the proximal and distal tracheal segments. The macaw recovered without complication and has had no recurrence of respiratory abnormalities 2 years after the second surgery. This report describes the longest total tracheal segment to be resected, followed by tracheal anastomosis, in a psittacine bird. The positive outcome in this case suggests that, when surgical therapy is elected, an aggressive approach is necessary to prevent recurrence of tracheal stenosis. In addition, macaws can recover well even after significant lengths of the trachea are resected
    corecore