arXiv Analytics

Sign in

arXiv:2403.09296 [cs.CV]AbstractReferencesReviewsResources

Select and Distill: Selective Dual-Teacher Knowledge Transfer for Continual Learning on Vision-Language Models

Yu-Chu Yu, Chi-Pin Huang, Jr-Jen Chen, Kai-Po Chang, Yung-Hsuan Lai, Fu-En Yang, Yu-Chiang Frank Wang

Published 2024-03-14Version 1

Large-scale vision-language models (VLMs) have shown a strong zero-shot generalization capability on unseen-domain data. However, when adapting pre-trained VLMs to a sequence of downstream tasks, they are prone to forgetting previously learned knowledge and degrade their zero-shot classification capability. To tackle this problem, we propose a unique Selective Dual-Teacher Knowledge Transfer framework that leverages the most recent fine-tuned and the original pre-trained VLMs as dual teachers to preserve the previously learned knowledge and zero-shot capabilities, respectively. With only access to an unlabeled reference dataset, our proposed framework performs a selective knowledge distillation mechanism by measuring the feature discrepancy from the dual teacher VLMs. Consequently, our selective dual-teacher knowledge distillation would mitigate catastrophic forgetting of previously learned knowledge while preserving the zero-shot capabilities from pre-trained VLMs. Through extensive experiments on benchmark datasets, we show that our proposed framework is favorable against state-of-the-art continual learning approaches for preventing catastrophic forgetting and zero-shot degradation.

Related articles: Most relevant | Search more
arXiv:2411.06764 [cs.CV] (Published 2024-11-11)
Multi-Stage Knowledge Integration of Vision-Language Models for Continual Learning
arXiv:2403.19137 [cs.CV] (Published 2024-03-28)
CLAP4CLIP: Continual Learning with Probabilistic Finetuning for Vision-Language Models
arXiv:2307.11386 [cs.CV] (Published 2023-07-21)
CLR: Channel-wise Lightweight Reprogramming for Continual Learning