arXiv Analytics

Sign in

arXiv:1501.03326 [stat.ML]AbstractReferencesReviewsResources

Unbiased Bayes for Big Data: Paths of Partial Posteriors

Heiko Strathmann, Dino Sejdinovic, Mark Girolami

Published 2015-01-14Version 1

Bayesian inference proceeds based on expectations of certain functions with respect to the posterior. Markov Chain Monte Carlo is a fundamental tool to compute these expectations. However, its feasibility is being challenged in the era of so called Big Data as all data needs to be processed in every iteration. Realising that such simulation is an unnecessarily hard problem if the goal is estimation, we construct a computationally scalable methodology that allows unbiased estimation of the required expectations -- without explicit simulation from the full posterior. The average computational complexity of our scheme is sub-linear in the size of the dataset and its variance is straightforward to control, leading to algorithms that are provably unbiased and naturally arrive at a desired error tolerance. We demonstrate the utility and generality of the methodology on a range of common statistical models applied to large scale benchmark datasets.

Related articles: Most relevant | Search more
arXiv:1910.06539 [stat.ML] (Published 2019-10-15)
Challenges in Bayesian inference via Markov chain Monte Carlo for neural networks
arXiv:2011.06043 [stat.ML] (Published 2020-11-11)
Clustering of Big Data with Mixed Features
arXiv:1509.02900 [stat.ML] (Published 2015-09-09)
Statistical Inference, Learning and Models in Big Data
Beate Franke et al.