arXiv:1807.01280 [cs.LG]AbstractReferencesReviewsResources
On the Computational Power of Online Gradient Descent
Vaggos Chatziafratis, Tim Roughgarden, Joshua R. Wang
Published 2018-07-03Version 1
We prove that the evolution of weight vectors in online gradient descent can encode arbitrary polynomial-space computations, even in the special case of soft-margin support vector machines. Our results imply that, under weak complexity-theoretic assumptions, it is impossible to reason efficiently about the fine-grained behavior of online gradient descent.
Related articles: Most relevant | Search more
arXiv:2006.09286 [cs.LG] (Published 2020-06-16)
On the Computational Power of Transformers and Its Implications in Sequence Modeling
arXiv:1802.04623 [cs.LG] (Published 2018-02-13)
Fast Rates for Online Gradient Descent Without Strong Convexity via Hoffman's Bound
arXiv:1504.02089 [cs.LG] (Published 2015-04-08)
The Computational Power of Optimization in Online Learning