arXiv Analytics

Sign in

arXiv:1811.03679 [stat.ML]AbstractReferencesReviewsResources

Practical Bayesian Learning of Neural Networks via Adaptive Subgradient Methods

Arnold Salas, Stefan Zohren, Stephen Roberts

Published 2018-11-08Version 1

We introduce a novel framework for the estimation of the posterior distribution of the weights of a neural network, based on a new probabilistic interpretation of adaptive subgradient algorithms such as AdaGrad and Adam. Having a confidence measure of the weights allows several shortcomings of neural networks to be addressed. In particular, the robustness of the network can be improved by performing weight pruning based on signal-to-noise ratios from the weight posterior distribution. Using the MNIST dataset, we demonstrate that the empirical performance of Badam, a particular instance of our framework based on Adam, is competitive in comparison to related Bayesian approaches such as Bayes By Backprop.

Comments: Manuscript under review by AISTATS 2019
Categories: stat.ML, cs.LG
Related articles: Most relevant | Search more
arXiv:2106.13682 [stat.ML] (Published 2021-06-25)
Prediction of Hereditary Cancers Using Neural Networks
arXiv:1901.03838 [stat.ML] (Published 2019-01-12)
Enhancing Explainability of Neural Networks through Architecture Constraints
arXiv:2505.09167 [stat.ML] (Published 2025-05-14)
Online Learning of Neural Networks