{ "id": "2007.04803", "version": "v1", "published": "2020-07-09T14:17:43.000Z", "updated": "2020-07-09T14:17:43.000Z", "title": "Online Approximate Bayesian learning", "authors": [ "Mathieu Gerber", "Randal Douc" ], "comment": "76 pages(including an Appendix of 43 pages), 3 figures, 2 tables", "categories": [ "stat.ML", "math.ST", "stat.CO", "stat.TH" ], "abstract": "We introduce in this work a new approach for online approximate Bayesian learning. The main idea of the proposed method is to approximate the sequence $(\\pi_t)_{t\\geq 1}$ of posterior distributions by a sequence $(\\tilde{\\pi}_t)_{t\\geq 1}$ which (i) can be estimated in an online fashion using sequential Monte Carlo methods and (ii) is shown to converge to the same distribution as the sequence $(\\pi_t)_{t\\geq 1}$, under weak assumptions on the statistical model at hand. In its simplest version, $(\\tilde{\\pi}_t)_{t\\geq 1}$ is the sequence of filtering distributions associated to a particular state-space model, which can therefore be approximated using a standard particle filter algorithm. We illustrate on several challenging examples the benefits of this approach for approximate Bayesian parameter inference, and with one real data example we show that its online predictive performance can significantly outperform that of stochastic gradient descent and streaming variational Bayes.", "revisions": [ { "version": "v1", "updated": "2020-07-09T14:17:43.000Z" } ], "analyses": { "keywords": [ "online approximate bayesian learning", "standard particle filter algorithm", "sequential monte carlo methods", "approximate bayesian parameter inference", "real data example" ], "note": { "typesetting": "TeX", "pages": 76, "language": "en", "license": "arXiv", "status": "editable" } } }