arXiv Analytics

Sign in

arXiv:2405.16644 [stat.ML]AbstractReferencesReviewsResources

Gaussian Approximation and Multiplier Bootstrap for Polyak-Ruppert Averaged Linear Stochastic Approximation with Applications to TD Learning

Sergey Samsonov, Eric Moulines, Qi-Man Shao, Zhuo-Song Zhang, Alexey Naumov

Published 2024-05-26Version 1

In this paper, we obtain the Berry-Esseen bound for multivariate normal approximation for the Polyak-Ruppert averaged iterates of the linear stochastic approximation (LSA) algorithm with decreasing step size. Our findings reveal that the fastest rate of normal approximation is achieved when setting the most aggressive step size $\alpha_{k} \asymp k^{-1/2}$. Moreover, we prove the non-asymptotic validity of the confidence intervals for parameter estimation with LSA based on multiplier bootstrap. This procedure updates the LSA estimate together with a set of randomly perturbed LSA estimates upon the arrival of subsequent observations. We illustrate our findings in the setting of temporal difference learning with linear function approximation.

Related articles: Most relevant | Search more
arXiv:1209.3079 [stat.ML] (Published 2012-09-14)
Signal Recovery in Unions of Subspaces with Applications to Compressive Imaging
arXiv:1806.08301 [stat.ML] (Published 2018-06-21)
Online Saddle Point Problem with Applications to Constrained Online Convex Optimization
arXiv:2102.04259 [stat.ML] (Published 2021-02-04)
Concentration of Non-Isotropic Random Tensors with Applications to Learning and Empirical Risk Minimization