arXiv Analytics

Sign in

arXiv:1905.05843 [cs.LG]AbstractReferencesReviewsResources

Task-Driven Data Verification via Gradient Descent

Siavash Golkar, Kyunghyun Cho

Published 2019-05-14Version 1

We introduce a novel algorithm for the detection of possible sample corruption such as mislabeled samples in a training dataset given a small clean validation set. We use a set of inclusion variables which determine whether or not any element of the noisy training set should be included in the training of a network. We compute these inclusion variables by optimizing the performance of the network on the clean validation set via "gradient descent on gradient descent" based learning. The inclusion variables as well as the network trained in such a way form the basis of our methods, which we call Corruption Detection via Gradient Descent (CDGD). This algorithm can be applied to any supervised machine learning task and is not limited to classification problems. We provide a quantitative comparison of these methods on synthetic and real world datasets.

Related articles: Most relevant | Search more
arXiv:2204.08809 [cs.LG] (Published 2022-04-19)
Making Progress Based on False Discoveries
arXiv:1903.11680 [cs.LG] (Published 2019-03-27)
Gradient Descent with Early Stopping is Provably Robust to Label Noise for Overparameterized Neural Networks
arXiv:2011.02402 [cs.LG] (Published 2020-11-04)
On the Convergence of Gradient Descent in GANs: MMD GAN As a Gradient Flow