arXiv Analytics

Sign in

arXiv:2102.08668 [math.PR]AbstractReferencesReviewsResources

Non-asymptotic approximations of neural networks by Gaussian processes

Ronen Eldan, Dan Mikulincer, Tselil Schramm

Published 2021-02-17Version 1

We study the extent to which wide neural networks may be approximated by Gaussian processes when initialized with random weights. It is a well-established fact that as the width of a network goes to infinity, its law converges to that of a Gaussian process. We make this quantitative by establishing explicit convergence rates for the central limit theorem in an infinite-dimensional functional space, metrized with a natural transportation distance. We identify two regimes of interest; when the activation function is polynomial, its degree determines the rate of convergence, while for non-polynomial activations, the rate is governed by the smoothness of the function.

Related articles: Most relevant | Search more
arXiv:2001.03405 [math.PR] (Published 2020-01-10)
Gaussian processes with Volterra kernels
arXiv:2502.06983 [math.PR] (Published 2025-02-10)
Riemann-Skorohod and Stratonovich integrals for Gaussian processes
arXiv:2412.10001 [math.PR] (Published 2024-12-13)
On the Markov transformation of Gaussian processes