arXiv Analytics

Sign in

arXiv:1708.01383 [cs.LG]AbstractReferencesReviewsResources

Convergence of Variance-Reduced Stochastic Learning under Random Reshuffling

Bicheng Ying, Kun Yuan, Ali H. Sayed

Published 2017-08-04Version 1

Several useful variance-reduced stochastic gradient algorithms, such as SVRG, SAGA, Finito, and SAG, have been proposed to minimize empirical risks with linear convergence properties to the exact minimizers. The existing convergence results assume uniform data sampling with replacement. However, it has been observed that random reshuffling can deliver superior performance. No formal proofs or guarantees of exact convergence exist for variance-reduced algorithms under random reshuffling. This paper resolves this open convergence issue and provides the first theoretical guarantee of linear convergence under random reshuffling for SAGA; the argument is also adaptable to other variance-reduced algorithms. Under random reshuffling, the paper further proposes a new amortized variance-reduced gradient (AVRG) algorithm with constant storage requirements compared to SAGA and with balanced gradient computations compared to SVRG. The balancing in computations are attained by amortizing the full gradient calculation across all iterations. AVRG is also shown analytically to converge linearly.

Related articles: Most relevant | Search more
arXiv:2104.12112 [cs.LG] (Published 2021-04-25)
On the Comparison between Cyclic Sampling and Random Reshuffling
arXiv:2205.03914 [cs.LG] (Published 2022-05-08)
Random Reshuffling with Variance Reduction: New Analysis and Better Rates
arXiv:2304.00459 [cs.LG] (Published 2023-04-02)
Fast Convergence of Random Reshuffling under Over-Parameterization and the Polyak-Łojasiewicz Condition