arXiv Analytics

Sign in

arXiv:2406.06903 [stat.ML]AbstractReferencesReviewsResources

On the Limitation of Kernel Dependence Maximization for Feature Selection

Keli Liu, Feng Ruan

Published 2024-06-11Version 1

A simple and intuitive method for feature selection consists of choosing the feature subset that maximizes a nonparametric measure of dependence between the response and the features. A popular proposal from the literature uses the Hilbert-Schmidt Independence Criterion (HSIC) as the nonparametric dependence measure. The rationale behind this approach to feature selection is that important features will exhibit a high dependence with the response and their inclusion in the set of selected features will increase the HSIC. Through counterexamples, we demonstrate that this rationale is flawed and that feature selection via HSIC maximization can miss critical features.

Related articles: Most relevant | Search more
arXiv:1809.01706 [stat.ML] (Published 2018-08-27)
A Limitation of V-Matrix based Methods
arXiv:2302.09930 [stat.ML] (Published 2023-02-20)
Nyström $M$-Hilbert-Schmidt Independence Criterion
arXiv:2106.08320 [stat.ML] (Published 2021-06-15)
Self-Supervised Learning with Kernel Dependence Maximization