arXiv Analytics

Sign in

arXiv:2106.10885 [cs.CV]AbstractReferencesReviewsResources

Knowledge Distillation via Instance-level Sequence Learning

Haoran Zhao, Xin Sun, Junyu Dong, Zihe Dong, Qiong Li

Published 2021-06-21Version 1

Recently, distillation approaches are suggested to extract general knowledge from a teacher network to guide a student network. Most of the existing methods transfer knowledge from the teacher network to the student via feeding the sequence of random mini-batches sampled uniformly from the data. Instead, we argue that the compact student network should be guided gradually using samples ordered in a meaningful sequence. Thus, it can bridge the gap of feature representation between the teacher and student network step by step. In this work, we provide a curriculum learning knowledge distillation framework via instance-level sequence learning. It employs the student network of the early epoch as a snapshot to create a curriculum for the student network's next training phase. We carry out extensive experiments on CIFAR-10, CIFAR-100, SVHN and CINIC-10 datasets. Compared with several state-of-the-art methods, our framework achieves the best performance with fewer iterations.

Related articles: Most relevant | Search more
arXiv:1710.09505 [cs.CV] (Published 2017-10-26)
Knowledge Projection for Deep Neural Networks
arXiv:2409.07694 [cs.CV] (Published 2024-09-12)
Learn from Balance: Rectifying Knowledge Transfer for Long-Tailed Scenarios
arXiv:2303.05073 [cs.CV] (Published 2023-03-09, updated 2023-08-15)
Learn More for Food Recognition via Progressive Self-Distillation