arXiv Analytics

Sign in

arXiv:2501.11773 [stat.ML]AbstractReferencesReviewsResources

Can Bayesian Neural Networks Make Confident Predictions?

Katharine Fisher, Youssef Marzouk

Published 2025-01-20Version 1

Bayesian inference promises a framework for principled uncertainty quantification of neural network predictions. Barriers to adoption include the difficulty of fully characterizing posterior distributions on network parameters and the interpretability of posterior predictive distributions. We demonstrate that under a discretized prior for the inner layer weights, we can exactly characterize the posterior predictive distribution as a Gaussian mixture. This setting allows us to define equivalence classes of network parameter values which produce the same likelihood (training error) and to relate the elements of these classes to the network's scaling regime -- defined via ratios of the training sample size, the size of each layer, and the number of final layer parameters. Of particular interest are distinct parameter realizations that map to low training error and yet correspond to distinct modes in the posterior predictive distribution. We identify settings that exhibit such predictive multimodality, and thus provide insight into the accuracy of unimodal posterior approximations. We also characterize the capacity of a model to "learn from data" by evaluating contraction of the posterior predictive in different scaling regimes.

Comments: Mathematics of Modern Machine Learning Workshop at NeurIPS 2024
Categories: stat.ML, cs.LG, math.ST, stat.TH
Related articles: Most relevant | Search more
arXiv:2309.16314 [stat.ML] (Published 2023-09-28)
A Primer on Bayesian Neural Networks: Review and Debates
arXiv:1903.07594 [stat.ML] (Published 2019-03-18)
Combining Model and Parameter Uncertainty in Bayesian Neural Networks
arXiv:2006.12024 [stat.ML] (Published 2020-06-22)
Bayesian Neural Networks: An Introduction and Survey