arXiv Analytics

Sign in

arXiv:2302.06354 [cs.LG]AbstractReferencesReviewsResources

SubTuning: Efficient Finetuning for Multi-Task Learning

Gal Kaplun, Andrey Gurevich, Tal Swisa, Mazor David, Shai Shalev-Shwartz, Eran Malach

Published 2023-02-13Version 1

Finetuning a pretrained model has become a standard approach for training neural networks on novel tasks, resulting in fast convergence and improved performance. In this work, we study an alternative finetuning method, where instead of finetuning all the weights of the network, we only train a carefully chosen subset of layers, keeping the rest of the weights frozen at their initial (pretrained) values. We demonstrate that \emph{subset finetuning} (or SubTuning) often achieves accuracy comparable to full finetuning of the model, and even surpasses the performance of full finetuning when training data is scarce. Therefore, SubTuning allows deploying new tasks at minimal computational cost, while enjoying the benefits of finetuning the entire model. This yields a simple and effective method for multi-task learning, where different tasks do not interfere with one another, and yet share most of the resources at inference time. We demonstrate the efficiency of SubTuning across multiple tasks, using different network architectures and pretraining methods.

Related articles: Most relevant | Search more
arXiv:2002.04792 [cs.LG] (Published 2020-02-12)
A Simple General Approach to Balance Task Difficulty in Multi-Task Learning
arXiv:1203.3536 [cs.LG] (Published 2012-03-15)
A Convex Formulation for Learning Task Relationships in Multi-Task Learning
arXiv:1707.08114 [cs.LG] (Published 2017-07-25)
A Survey on Multi-Task Learning