arXiv Analytics

Sign in

arXiv:1712.03428 [cs.LG]AbstractReferencesReviewsResources

Cost-Sensitive Approach to Batch Size Adaptation for Gradient Descent

Matteo Pirotta, Marcello Restelli

Published 2017-12-09Version 1

In this paper, we propose a novel approach to automatically determine the batch size in stochastic gradient descent methods. The choice of the batch size induces a trade-off between the accuracy of the gradient estimate and the cost in terms of samples of each update. We propose to determine the batch size by optimizing the ratio between a lower bound to a linear or quadratic Taylor approximation of the expected improvement and the number of samples used to estimate the gradient. The performance of the proposed approach is empirically compared with related methods on popular classification tasks. The work was presented at the NIPS workshop on Optimizing the Optimizers. Barcelona, Spain, 2016.

Comments: Presented at the NIPS workshop on Optimizing the Optimizers. Barcelona, Spain, 2016
Categories: cs.LG, stat.ML
Related articles: Most relevant | Search more
arXiv:1512.02970 [cs.LG] (Published 2015-12-09)
Scaling Up Distributed Stochastic Gradient Descent Using Variance Reduction
arXiv:1703.07948 [cs.LG] (Published 2017-03-23)
Fast Stochastic Variance Reduced Gradient Method with Momentum Acceleration for Machine Learning
arXiv:1612.05086 [cs.LG] (Published 2016-12-15)
Coupling Adaptive Batch Sizes with Learning Rates