arXiv Analytics

Sign in

arXiv:2101.04097 [stat.ML]AbstractReferencesReviewsResources

Correlated Weights in Infinite Limits of Deep Convolutional Neural Networks

Adrià Garriga-Alonso, Mark van der Wilk

Published 2021-01-11Version 1

Infinite width limits of deep neural networks often have tractable forms. They have been used to analyse the behaviour of finite networks, as well as being useful methods in their own right. When investigating infinitely wide CNNs it was observed that the correlations arising from spatial weight sharing disappear in the infinite limit. This is undesirable, as spatial correlation is the main motivation behind CNNs. We show that the loss of this property is not a consequence of the infinite limit, but rather of choosing an independent weight prior. Correlating the weights maintains the correlations in the activations. Varying the amount of correlation interpolates between independent-weight limits and mean-pooling. Empirical evaluation of the infinitely wide network shows that optimal performance is achieved between the extremes, indicating that correlations can be useful.

Comments: Presented at 3rd Symposium on Advances in Approximate Bayesian Inference
Categories: stat.ML, cs.LG
Related articles: Most relevant | Search more
arXiv:1909.11396 [stat.ML] (Published 2019-09-25)
Information Plane Analysis of Deep Neural Networks via Matrix-Based Renyi's Entropy and Tensor Kernels
arXiv:1611.07429 [stat.ML] (Published 2016-11-22)
TreeView: Peeking into Deep Neural Networks Via Feature-Space Partitioning
arXiv:1802.07714 [stat.ML] (Published 2018-02-21)
Detecting Learning vs Memorization in Deep Neural Networks using Shared Structure Validation Sets