arXiv:1806.00468 [cs.LG]AbstractReferencesReviewsResources
Implicit Bias of Gradient Descent on Linear Convolutional Networks
Suriya Gunasekar, Jason Lee, Daniel Soudry, Nathan Srebro
Published 2018-06-01Version 1
We show that gradient descent on full-width linear convolutional networks of depth $L$ converges to a linear predictor related to the $\ell_{2/L}$ bridge penalty in the frequency domain. This is in contrast to linearly fully connected networks, where gradient descent converges to the hard margin linear support vector machine solution, regardless of depth.
Related articles: Most relevant | Search more
arXiv:2202.04302 [cs.LG] (Published 2022-02-09)
On the Implicit Bias of Gradient Descent for Temporal Extrapolation
arXiv:2204.08809 [cs.LG] (Published 2022-04-19)
Making Progress Based on False Discoveries
arXiv:2006.10925 [cs.LG] (Published 2020-06-19)
Gradient Descent in RKHS with Importance Labeling