arXiv Analytics

Sign in

arXiv:2004.12289 [cs.LG]AbstractReferencesReviewsResources

Deep k-NN for Noisy Labels

Dara Bahri, Heinrich Jiang, Maya Gupta

Published 2020-04-26Version 1

Modern machine learning models are often trained on examples with noisy labels that hurt performance and are hard to identify. In this paper, we provide an empirical study showing that a simple $k$-nearest neighbor-based filtering approach on the logit layer of a preliminary model can remove mislabeled training data and produce more accurate models than many recently proposed methods. We also provide new statistical guarantees into its efficacy.

Comments: Full paper (including supplemental) can be found at https://github.com/dbahri/deepknn
Categories: cs.LG, cs.AI, stat.ML
Related articles: Most relevant | Search more
arXiv:1910.03231 [cs.LG] (Published 2019-10-08)
Peer Loss Functions: Learning from Noisy Labels without Knowing Noise Rates
arXiv:2106.00274 [cs.LG] (Published 2021-06-01)
Analysis of classifiers robust to noisy labels
arXiv:2208.12807 [cs.LG] (Published 2022-08-25)
Towards Federated Learning against Noisy Labels via Local Self-Regularization