arXiv Analytics

Sign in

arXiv:1808.01174 [cs.LG]AbstractReferencesReviewsResources

Generalization Error in Deep Learning

Daniel Jakubovitz, Raja Giryes, Miguel R. D. Rodrigues

Published 2018-08-03Version 1

Deep learning models have lately shown great performance in various fields such as computer vision, speech recognition, speech translation, and natural language processing. However, alongside their state-of-the-art performance, it is still generally unclear what is the source of their generalization ability. Thus, an important question is what makes deep neural networks able to generalize well from the training set to new data. In this article, we provide an overview of the existing theory and bounds for the characterization of the generalization error of deep neural networks, combining both classical and more recent theoretical and empirical results.

Related articles: Most relevant | Search more
arXiv:2202.09275 [cs.LG] (Published 2022-02-18)
Rethinking Pareto Frontier for Performance Evaluation of Deep Neural Networks
arXiv:1908.10714 [cs.LG] (Published 2019-08-22)
Automated Architecture Design for Deep Neural Networks
arXiv:2011.06796 [cs.LG] (Published 2020-11-13)
Wisdom of the Ensemble: Improving Consistency of Deep Learning Models
Lijing Wang et al.