arXiv Analytics

Sign in

arXiv:1904.01802 [cs.CV]AbstractReferencesReviewsResources

Correlation Congruence for Knowledge Distillation

Baoyun Peng, Xiao Jin, Jiaheng Liu, Shunfeng Zhou, Yichao Wu, Yu Liu, Dongsheng Li, Zhaoning Zhang

Published 2019-04-03Version 1

Most teacher-student frameworks based on knowledge distillation (KD) depend on a strong congruent constraint on instance level. However, they usually ignore the correlation between multiple instances, which is also valuable for knowledge transfer. In this work, we propose a new framework named correlation congruence for knowledge distillation (CCKD), which transfers not only the instance-level information, but also the correlation between instances. Furthermore, a generalized kernel method based on Taylor series expansion is proposed to better capture the correlation between instances. Empirical experiments and ablation studies on image classification tasks (including CIFAR-100, ImageNet-1K) and metric learning tasks (including ReID and Face Recognition) show that the proposed CCKD substantially outperforms the original KD and achieves state-of-the-art accuracy compared with other SOTA KD-based methods. The CCKD can be easily deployed in the majority of the teacher-student framework such as KD and hint-based learning methods.

Related articles: Most relevant | Search more
arXiv:2304.06619 [cs.CV] (Published 2023-04-13)
Class-Incremental Learning of Plant and Disease Detection: Growing Branches with Knowledge Distillation
arXiv:2305.07586 [cs.CV] (Published 2023-05-12)
Knowledge distillation with Segment Anything (SAM) model for Planetary Geological Mapping
arXiv:1907.09643 [cs.CV] (Published 2019-07-23)
Highlight Every Step: Knowledge Distillation via Collaborative Teaching