{ "id": "2008.08400", "version": "v1", "published": "2020-08-19T12:35:55.000Z", "updated": "2020-08-19T12:35:55.000Z", "title": "Improving predictions of Bayesian neural networks via local linearization", "authors": [ "Alexander Immer", "Maciej Korzepa", "Matthias Bauer" ], "comment": "ICML 2020 Workshop on Uncertainty and Robustness in Deep Learning", "categories": [ "stat.ML", "cs.LG" ], "abstract": "In this paper we argue that in Bayesian deep learning, the frequently utilized generalized Gauss-Newton (GGN) approximation should be understood as a modification of the underlying probabilistic model and should be considered separately from further approximate inference techniques. Applying the GGN approximation turns a BNN into a locally linearized generalized linear model or, equivalently, a Gaussian process. Because we then use this linearized model for inference, we should also predict using this modified likelihood rather than the original BNN likelihood. This formulation extends previous results to general likelihoods and alleviates underfitting behaviour observed e.g. by Ritter et al. (2018). We demonstrate our approach on several UCI classification datasets as well as CIFAR10.", "revisions": [ { "version": "v1", "updated": "2020-08-19T12:35:55.000Z" } ], "analyses": { "keywords": [ "bayesian neural networks", "local linearization", "improving predictions", "linearized generalized linear model", "approximate inference techniques" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }