arXiv Analytics

Sign in

arXiv:1911.05275 [cs.LG]AbstractReferencesReviewsResources

Learning from a Teacher using Unlabeled Data

Gaurav Menghani, Sujith Ravi

Published 2019-11-13Version 1

Knowledge distillation is a widely used technique for model compression. We posit that the teacher model used in a distillation setup, captures relationships between classes, that extend beyond the original dataset. We empirically show that a teacher model can transfer this knowledge to a student model even on an {\it out-of-distribution} dataset. Using this approach, we show promising results on MNIST, CIFAR-10, and Caltech-256 datasets using unlabeled image data from different sources. Our results are encouraging and help shed further light from the perspective of understanding knowledge distillation and utilizing unlabeled data to improve model quality.

Related articles: Most relevant | Search more
arXiv:1203.3495 [cs.LG] (Published 2012-03-15)
Parameter-Free Spectral Kernel Learning
arXiv:1809.03207 [cs.LG] (Published 2018-09-10)
Beyond the Selected Completely At Random Assumption for Learning from Positive and Unlabeled Data
arXiv:1809.05710 [cs.LG] (Published 2018-09-15)
Alternate Estimation of a Classifier and the Class-Prior from Positive and Unlabeled Data