arXiv Analytics

Sign in

arXiv:2112.01401 [cs.LG]AbstractReferencesReviewsResources

Newton methods based convolution neural networks using parallel processing

Ujjwal Thakur, Anuj Sharma

Published 2021-12-02, updated 2022-09-09Version 2

Training of convolutional neural networks is a high dimensional and a non-convex optimization problem. At present, it is inefficient in situations where parametric learning rates can not be confidently set. Some past works have introduced Newton methods for training deep neural networks. Newton methods for convolutional neural networks involve complicated operations. Finding the Hessian matrix in second-order methods becomes very complex as we mainly use the finite differences method with the image data. Newton methods for convolutional neural networks deals with this by using the sub-sampled Hessian Newton methods. In this paper, we have used the complete data instead of the sub-sampled methods that only handle partial data at a time. Further, we have used parallel processing instead of serial processing in mini-batch computations. The results obtained using parallel processing in this study, outperform the time taken by the previous approach.

Related articles: Most relevant | Search more
arXiv:2107.11045 [cs.LG] (Published 2021-07-23)
Ensemble of Convolution Neural Networks on Heterogeneous Signals for Sleep Stage Scoring
arXiv:2205.04064 [cs.LG] (Published 2022-05-09)
Augmentations: An Insight into their Effectiveness on Convolution Neural Networks
arXiv:2210.09147 [cs.LG] (Published 2022-10-17, updated 2022-12-02)
PARTIME: Scalable and Parallel Processing Over Time with Deep Neural Networks