arXiv Analytics

Sign in

arXiv:1809.05606 [cs.LG]AbstractReferencesReviewsResources

Non-iterative recomputation of dense layers for performance improvement of DCNN

Yimin Yang, Q. M. Jonathan Wu, Xiexing Feng, Thangarajah Akilan

Published 2018-09-14Version 1

An iterative method of learning has become a paradigm for training deep convolutional neural networks (DCNN). However, utilizing a non-iterative learning strategy can accelerate the training process of the DCNN and surprisingly such approach has been rarely explored by the deep learning (DL) community. It motivates this paper to introduce a non-iterative learning strategy that eliminates the backpropagation (BP) at the top dense or fully connected (FC) layers of DCNN, resulting in, lower training time and higher performance. The proposed method exploits the Moore-Penrose Inverse to pull back the current residual error to each FC layer, generating well-generalized features. Then using the recomputed features, i.e., the new generalized features the weights of each FC layer is computed according to the Moore-Penrose Inverse. We evaluate the proposed approach on six widely accepted object recognition benchmark datasets: Scene-15, CIFAR-10, CIFAR-100, SUN-397, Places365, and ImageNet. The experimental results show that the proposed method obtains significant improvements over 30 state-of-the-art methods. Interestingly, it also indicates that any DCNN with the proposed method can provide better performance than the same network with its original training based on BP.

Related articles: Most relevant | Search more
arXiv:1902.05967 [cs.LG] (Published 2019-02-15)
Parameter Efficient Training of Deep Convolutional Neural Networks by Dynamic Sparse Reparameterization
arXiv:1809.09399 [cs.LG] (Published 2018-09-25)
Non-Iterative Knowledge Fusion in Deep Convolutional Neural Networks
arXiv:1901.08624 [cs.LG] (Published 2019-01-24)
AutoShuffleNet: Learning Permutation Matrices via an Exact Lipschitz Continuous Penalty in Deep Convolutional Neural Networks