arXiv Analytics

Sign in

arXiv:1504.06825 [cs.LG]AbstractReferencesReviewsResources

Comparison of Training Methods for Deep Neural Networks

Patrick O. Glauner

Published 2015-04-26Version 1

This report describes the difficulties of training neural networks and in particular deep neural networks. It then provides a literature review of training methods for deep neural networks, with a focus on pre-training. It focuses on Deep Belief Networks composed of Restricted Boltzmann Machines and Stacked Autoencoders and provides an outreach on further and alternative approaches. It also includes related practical recommendations from the literature on training them. In the second part, initial experiments using some of the covered methods are performed on two databases. In particular, experiments are performed on the MNIST hand-written digit dataset and on facial emotion data from a Kaggle competition. The results are discussed in the context of results reported in other research papers. An error rate lower than the best contribution to the Kaggle competition is achieved using an optimized Stacked Autoencoder.

Related articles: Most relevant | Search more
arXiv:2010.09458 [cs.LG] (Published 2020-10-15)
Review and Comparison of Commonly Used Activation Functions for Deep Neural Networks
arXiv:2307.02973 [cs.LG] (Published 2023-07-06)
Pruning vs Quantization: Which is Better?
arXiv:2007.14166 [cs.LG] (Published 2020-07-28)
A Comparison of Optimization Algorithms for Deep Learning