arXiv Analytics

Sign in

arXiv:1905.01422 [cs.LG]AbstractReferencesReviewsResources

NAMSG: An Efficient Method For Training Neural Networks

Yushu Chen, Hao Jing, Wenlai Zhao, Zhiqiang Liu, Liang Qiao, Wei Xue, Haohuan Fu, Guangwen Yang

Published 2019-05-04Version 1

We introduce NAMSG, an adaptive first-order algorithm for training neural networks. The method is efficient in computation and memory, and straightforward to implement. It computes the gradients at configurable remote observation points, in order to expedite the convergence by adjusting the step size for directions with different curvatures, in the stochastic setting. It also scales the updating vector elementwise by a nonincreasing preconditioner, to take the advantages of AMSGRAD. We analyze the convergence properties for both convex and nonconvex problems, by modeling the training process as a dynamic system, and provide a guideline to select the observation distance without grid search. We also propose a datadependent regret bound, which guarantees the convergence in the convex setting. Experiments demonstrate that NAMSG works well in practice and compares favorably to popular adaptive methods, such as ADAM, NADAM, and AMSGRAD.

Related articles: Most relevant | Search more
arXiv:1909.03172 [cs.LG] (Published 2019-09-07)
Towards Understanding the Importance of Noise in Training Neural Networks
arXiv:1906.05661 [cs.LG] (Published 2019-06-13)
Training Neural Networks for and by Interpolation
arXiv:1905.05894 [cs.LG] (Published 2019-05-15)
Online Normalization for Training Neural Networks