arXiv:2004.12289 [cs.LG]AbstractReferencesReviewsResources
Deep k-NN for Noisy Labels
Dara Bahri, Heinrich Jiang, Maya Gupta
Published 2020-04-26Version 1
Modern machine learning models are often trained on examples with noisy labels that hurt performance and are hard to identify. In this paper, we provide an empirical study showing that a simple $k$-nearest neighbor-based filtering approach on the logit layer of a preliminary model can remove mislabeled training data and produce more accurate models than many recently proposed methods. We also provide new statistical guarantees into its efficacy.
Comments: Full paper (including supplemental) can be found at https://github.com/dbahri/deepknn
Keywords: noisy labels, deep k-nn, modern machine learning models, remove mislabeled training data, nearest neighbor-based filtering approach
Tags: github project
Related articles: Most relevant | Search more
arXiv:1910.03231 [cs.LG] (Published 2019-10-08)
Peer Loss Functions: Learning from Noisy Labels without Knowing Noise Rates
arXiv:2106.00274 [cs.LG] (Published 2021-06-01)
Analysis of classifiers robust to noisy labels
arXiv:2208.12807 [cs.LG] (Published 2022-08-25)
Towards Federated Learning against Noisy Labels via Local Self-Regularization