arXiv Analytics

Sign in

arXiv:1811.07006 [cs.LG]AbstractReferencesReviewsResources

Projected BNNs: Avoiding weight-space pathologies by learning latent representations of neural network weights

Melanie F. Pradier, Weiwei Pan, Jiayu Yao, Soumya Ghosh, Finale Doshi-velez

Published 2018-11-16, updated 2018-12-03Version 2

While modern neural networks are making remarkable gains in terms of predictive accuracy, characterizing uncertainty over the parameters of these models (in a Bayesian setting) is challenging because of the high-dimensionality of the network parameter space and the correlations between these parameters. In this paper, we introduce a novel framework for variational inference for Bayesian neural networks that (1) encodes complex distributions in high-dimensional parameter space with representations in a low-dimensional latent space and (2) performs inference efficiently on the low-dimensional representations. Across a large array of synthetic and real-world datasets, we show that our method improves uncertainty characterization and model generalization when compared with methods that work directly in the parameter space.

Related articles: Most relevant | Search more
arXiv:2412.05175 [cs.LG] (Published 2024-12-06)
Variational Encoder-Decoders for Learning Latent Representations of Physical Systems
arXiv:2002.07933 [cs.LG] (Published 2020-02-19)
Improving Generalization by Controlling Label-Noise Information in Neural Network Weights
arXiv:2302.01976 [cs.LG] (Published 2023-02-03)
SPARLING: Learning Latent Representations with Extremely Sparse Activations