arXiv Analytics

Sign in

arXiv:2410.17796 [cs.LG]AbstractReferencesReviewsResources

A Comprehensive Analysis on the Learning Curve in Kernel Ridge Regression

Tin Sum Cheng, Aurelien Lucchi, Anastasis Kratsios, David Belius

Published 2024-10-23Version 1

This paper conducts a comprehensive study of the learning curves of kernel ridge regression (KRR) under minimal assumptions. Our contributions are three-fold: 1) we analyze the role of key properties of the kernel, such as its spectral eigen-decay, the characteristics of the eigenfunctions, and the smoothness of the kernel; 2) we demonstrate the validity of the Gaussian Equivalent Property (GEP), which states that the generalization performance of KRR remains the same when the whitened features are replaced by standard Gaussian vectors, thereby shedding light on the success of previous analyzes under the Gaussian Design Assumption; 3) we derive novel bounds that improve over existing bounds across a broad range of setting such as (in)dependent feature vectors and various combinations of eigen-decay rates in the over/underparameterized regimes.

Related articles: Most relevant | Search more
arXiv:2401.01270 [cs.LG] (Published 2024-01-02)
Optimal Rates of Kernel Ridge Regression under Source Condition in Large Dimensions
arXiv:2312.05885 [cs.LG] (Published 2023-12-10)
Adaptive Parameter Selection for Kernel Ridge Regression
arXiv:2309.13337 [cs.LG] (Published 2023-09-23)
On the Asymptotic Learning Curves of Kernel Ridge Regression under Power-law Decay