10 research outputs found
Minimax lower bounds for function estimation on graphs
We study minimax lower bounds for function estimation problems on large graph
when the target function is smoothly varying over the graph. We derive minimax
rates in the context of regression and classification problems on graphs that
satisfy an asymptotic shape assumption and with a smoothness condition on the
target function, both formulated in terms of the graph Laplacian
Optimality of Poisson processes intensity learning with Gaussian processes
In this paper we provide theoretical support for the so-called "Sigmoidal
Gaussian Cox Process" approach to learning the intensity of an inhomogeneous
Poisson process on a -dimensional domain. This method was proposed by Adams,
Murray and MacKay (ICML, 2009), who developed a tractable computational
approach and showed in simulation and real data experiments that it can work
quite satisfactorily. The results presented in the present paper provide
theoretical underpinning of the method. In particular, we show how to tune the
priors on the hyper parameters of the model in order for the procedure to
automatically adapt to the degree of smoothness of the unknown intensity and to
achieve optimal convergence rates
Safe-Bayesian Generalized Linear Regression
We study generalized Bayesian inference under misspecification, i.e. when the
model is 'wrong but useful'. Generalized Bayes equips the likelihood with a
learning rate . We show that for generalized linear models (GLMs),
-generalized Bayes concentrates around the best approximation of the
truth within the model for specific , even under severely
misspecified noise, as long as the tails of the true distribution are
exponential. We derive MCMC samplers for generalized Bayesian lasso and
logistic regression and give examples of both simulated and real-world data in
which generalized Bayes substantially outperforms standard Bayes.Comment: Final version. Accepted to AISTATS 202
Safe-Bayesian Generalized Linear Regression
We study generalized Bayesian inference under misspecification, i.e. when the model is âwrong but usefulâ. Generalized Bayes equips the likelihood with a learning rate η. We show that for generalized linear models (GLMs), η-generalized Bayes concentrates around the best approximation of the truth within the model for specific ηeq1, even under severely misspecified noise, as long as the tails of the true distribution are exponential. We derive MCMC samplers for generalized Bayesian lasso and logistic regression and give examples of both simulated and real-world data in which generalized Bayes substantially outperforms standard Bayes
Precise small deviations in L 2 of some Gaussian processes appearing in the regression context
We find precise small deviation asymptotics with respect to the Hilbert norm for some special Gaussian processes connected to two regression schemes studied by MacNeill and his coauthors. In addition, we also obtain precise small deviation asymptotics for the detrended Brownian motion and detrended Slepian process
Safe Bayesian Linear Regression
We study generalized Bayesian inference under misspecification, i.e. when the model is `wrong but useful'. Generalized Bayes equips the likelihood with a learning rate η. We show that for generalized linear models (GLMs), η-generalized Bayes concentrates around the best approximation of the truth within the model for specific ηâ 1, even under severely misspecified noise, as long as the tails of the true distribution are exponential. We then derive MCMC samplers for generalized Bayesian lasso and logistic regression, and give examples of both simulated and real-world data in which generalized Bayes outperforms standard Bayes by a vast margin
Safe-Bayesian Generalized Linear Regression
We study generalized Bayesian inference under misspecification, i.e. when the model is âwrong but usefulâ. Generalized Bayes equips the likelihood with a learning rate η. We show that for generalized linear models (GLMs), η-generalized Bayes concentrates around the best approximation of the truth within the model for specific ηeq1, even under severely misspecified noise, as long as the tails of the true distribution are exponential. We derive MCMC samplers for generalized Bayesian lasso and logistic regression and give examples of both simulated and real-world data in which generalized Bayes substantially outperforms standard Bayes
Safe Bayesian Linear Regression
We study generalized Bayesian inference under misspecification, i.e. when the model is `wrong but useful'. Generalized Bayes equips the likelihood with a learning rate η. We show that for generalized linear models (GLMs), η-generalized Bayes concentrates around the best approximation of the truth within the model for specific ηâ 1, even under severely misspecified noise, as long as the tails of the true distribution are exponential. We then derive MCMC samplers for generalized Bayesian lasso and logistic regression, and give examples of both simulated and real-world data in which generalized Bayes outperforms standard Bayes by a vast margin