arXiv Analytics

Sign in

arXiv:2202.04925 [cond-mat.dis-nn]AbstractReferencesReviewsResources

Decomposing neural networks as mappings of correlation functions

Kirsten Fischer, Alexandre René, Christian Keup, Moritz Layer, David Dahmen, Moritz Helias

Published 2022-02-10Version 1

Understanding the functional principles of information processing in deep neural networks continues to be a challenge, in particular for networks with trained and thus non-random weights. To address this issue, we study the mapping between probability distributions implemented by a deep feed-forward network. We characterize this mapping as an iterated transformation of distributions, where the non-linearity in each layer transfers information between different orders of correlation functions. This allows us to identify essential statistics in the data, as well as different information representations that can be used by neural networks. Applied to an XOR task and to MNIST, we show that correlations up to second order predominantly capture the information processing in the internal layers, while the input layer also extracts higher-order correlations from the data. This analysis provides a quantitative and explainable perspective on classification.

Related articles: Most relevant | Search more
arXiv:cond-mat/9901110 (Published 1999-01-12, updated 1999-03-09)
Wigner-Dyson Statistics from the Replica Method
arXiv:1910.00631 [cond-mat.dis-nn] (Published 2019-10-01)
The correlation functions of certain random antiferromagnetic spin-1/2 critical chains
arXiv:cond-mat/0403625 (Published 2004-03-25, updated 2004-04-30)
Griffiths inequalities for the Gaussian spin glass