arXiv Analytics

Sign in

arXiv:1802.02219 [stat.ML]AbstractReferencesReviewsResources

Scalable Meta-Learning for Bayesian Optimization

Matthias Feurer, Benjamin Letham, Eytan Bakshy

Published 2018-02-06Version 1

Bayesian optimization has become a standard technique for hyperparameter optimization, including data-intensive models such as deep neural networks that may take days or weeks to train. We consider the setting where previous optimization runs are available, and we wish to use their results to warm-start a new optimization run. We develop an ensemble model that can incorporate the results of past optimization runs, while avoiding the poor scaling that comes with putting all results into a single Gaussian process model. The ensemble combines models from past runs according to estimates of their generalization performance on the current optimization. Results from a large collection of hyperparameter optimization benchmark problems and from optimization of a production computer vision platform at Facebook show that the ensemble can substantially reduce the time it takes to obtain near-optimal configurations, and is useful for warm-starting expensive searches or running quick re-optimizations.

Related articles: Most relevant | Search more
arXiv:1709.07842 [stat.ML] (Published 2017-09-22)
Bayesian Optimization for Parameter Tuning of the XOR Neural Network
arXiv:1910.05484 [stat.ML] (Published 2019-10-12)
Bayesian Optimization using Pseudo-Points
arXiv:2301.05983 [stat.ML] (Published 2023-01-14)
On the role of Model Uncertainties in Bayesian Optimization