arXiv Analytics

Sign in

arXiv:2003.11342 [cs.CV]AbstractReferencesReviewsResources

Circumventing Outliers of AutoAugment with Knowledge Distillation

Longhui Wei, An Xiao, Lingxi Xie, Xin Chen, Xiaopeng Zhang, Qi Tian

Published 2020-03-25Version 1

AutoAugment has been a powerful algorithm that improves the accuracy of many vision tasks, yet it is sensitive to the operator space as well as hyper-parameters, and an improper setting may degenerate network optimization. This paper delves deep into the working mechanism, and reveals that AutoAugment may remove part of discriminative information from the training image and so insisting on the ground-truth label is no longer the best option. To relieve the inaccuracy of supervision, we make use of knowledge distillation that refers to the output of a teacher model to guide network training. Experiments are performed in standard image classification benchmarks, and demonstrate the effectiveness of our approach in suppressing noise of data augmentation and stabilizing training. Upon the cooperation of knowledge distillation and AutoAugment, we claim the new state-of-the-art on ImageNet classification with a top-1 accuracy of 85.8%.

Related articles: Most relevant | Search more
arXiv:2006.16589 [cs.CV] (Published 2020-06-30)
On the Demystification of Knowledge Distillation: A Residual Network Perspective
arXiv:2303.08360 [cs.CV] (Published 2023-03-15)
Knowledge Distillation from Single to Multi Labels: an Empirical Study
arXiv:1904.01802 [cs.CV] (Published 2019-04-03)
Correlation Congruence for Knowledge Distillation
Baoyun Peng et al.