arXiv Analytics

Sign in

arXiv:1402.5836 [stat.ML]AbstractReferencesReviewsResources

Avoiding pathologies in very deep networks

David Duvenaud, Oren Rippel, Ryan P. Adams, Zoubin Ghahramani

Published 2014-02-24, updated 2014-09-14Version 2

Choosing appropriate architectures and regularization strategies for deep networks is crucial to good predictive performance. To shed light on this problem, we analyze the analogous problem of constructing useful priors on compositions of functions. Specifically, we study the deep Gaussian process, a type of infinitely-wide, deep neural network. We show that in standard architectures, the representational capacity of the network tends to capture fewer degrees of freedom as the number of layers increases, retaining only a single degree of freedom in the limit. We propose an alternate network architecture which does not suffer from this pathology. We also examine deep covariance functions, obtained by composing infinitely many feature transforms. Lastly, we characterize the class of models obtained by performing dropout on Gaussian processes.

Comments: 20 pages, 14 figures. Appeared in AISTATS 2014. This version has many minor fixes, and nicer figures
Categories: stat.ML, cs.LG
Related articles: Most relevant | Search more
arXiv:1903.09215 [stat.ML] (Published 2019-03-21)
Empirical confidence estimates for classification by deep neural networks
arXiv:2310.01683 [stat.ML] (Published 2023-10-02)
Commutative Width and Depth Scaling in Deep Neural Networks
arXiv:1607.00485 [stat.ML] (Published 2016-07-02)
Group Sparse Regularization for Deep Neural Networks