arXiv Analytics

Sign in

arXiv:2206.08476 [cs.LG]AbstractReferencesReviewsResources

Zero-Shot AutoML with Pretrained Models

Ekrem Öztürk, Fabio Ferreira, Hadi S. Jomaa, Lars Schmidt-Thieme, Josif Grabocka, Frank Hutter

Published 2022-06-16Version 1

Given a new dataset D and a low compute budget, how should we choose a pre-trained model to fine-tune to D, and set the fine-tuning hyperparameters without risking overfitting, particularly if D is small? Here, we extend automated machine learning (AutoML) to best make these choices. Our domain-independent meta-learning approach learns a zero-shot surrogate model which, at test time, allows to select the right deep learning (DL) pipeline (including the pre-trained model and fine-tuning hyperparameters) for a new dataset D given only trivial meta-features describing D such as image resolution or the number of classes. To train this zero-shot model, we collect performance data for many DL pipelines on a large collection of datasets and meta-train on this data to minimize a pairwise ranking objective. We evaluate our approach under the strict time limit of the vision track of the ChaLearn AutoDL challenge benchmark, clearly outperforming all challenge contenders.

Journal: International Conference on Machine Learning 2022
Categories: cs.LG, cs.AI, cs.CV
Related articles: Most relevant | Search more
arXiv:2301.09820 [cs.LG] (Published 2023-01-24)
A Stability Analysis of Fine-Tuning a Pre-Trained Model
arXiv:2203.10789 [cs.LG] (Published 2022-03-21)
Domain Generalization by Mutual-Information Regularization with Pre-trained Models
arXiv:2408.09310 [cs.LG] (Published 2024-08-17)
Narrowing the Focus: Learned Optimizers for Pretrained Models