arXiv Analytics

Sign in

arXiv:2207.01234 [cs.LG]AbstractReferencesReviewsResources

Look beyond labels: Incorporating functional summary information in Bayesian neural networks

Vishnu Raj, Tianyu Cui, Markus Heinonen, Pekka Marttinen

Published 2022-07-04Version 1

Bayesian deep learning offers a principled approach to train neural networks that accounts for both aleatoric and epistemic uncertainty. In variational inference, priors are often specified over the weight parameters, but they do not capture the true prior knowledge in large and complex neural network architectures. We present a simple approach to incorporate summary information about the predicted probability (such as sigmoid or softmax score) outputs in Bayesian neural networks (BNNs). The available summary information is incorporated as augmented data and modeled with a Dirichlet process, and we derive the corresponding \emph{Summary Evidence Lower BOund}. We show how the method can inform the model about task difficulty or class imbalance. Extensive empirical experiments show that, with negligible computational overhead, the proposed method yields a BNN with a better calibration of uncertainty.

Related articles: Most relevant | Search more
arXiv:2110.06020 [cs.LG] (Published 2021-10-12, updated 2022-02-21)
On out-of-distribution detection with Bayesian neural networks
arXiv:2304.01762 [cs.LG] (Published 2023-04-04)
Incorporating Unlabelled Data into Bayesian Neural Networks
arXiv:1605.09674 [cs.LG] (Published 2016-05-31)
Curiosity-driven Exploration in Deep Reinforcement Learning via Bayesian Neural Networks