arXiv Analytics

Sign in

arXiv:1705.07477 [cs.LG]AbstractReferencesReviewsResources

Statistical inference using SGD

Tianyang Li, Liu Liu, Anastasios Kyrillidis, Constantine Caramanis

Published 2017-05-21Version 1

We present a novel method for frequentist statistical inference in $M$-estimation problems, based on stochastic gradient descent (SGD) with a fixed step size: we demonstrate that the average of such SGD sequences can be used for statistical inference, after proper scaling. An intuitive analysis using the Ornstein-Uhlenbeck process suggests that such averages are asymptotically normal. From a practical perspective, our SGD-based inference procedure is a first order method, and is well-suited for large scale problems. To show its merits, we apply it to both synthetic and real datasets, and demonstrate that its accuracy is comparable to classical statistical methods, while requiring potentially far less computation.

Related articles: Most relevant | Search more
arXiv:1509.01240 [cs.LG] (Published 2015-09-03)
Train faster, generalize better: Stability of stochastic gradient descent
arXiv:1212.1824 [cs.LG] (Published 2012-12-08, updated 2012-12-28)
Stochastic Gradient Descent for Non-smooth Optimization: Convergence Results and Optimal Averaging Schemes
arXiv:2006.04760 [cs.LG] (Published 2020-06-08)
Outlier Detection Using a Novel method: Quantum Clustering