arXiv Analytics

Sign in

arXiv:1904.05658 [cs.LG]AbstractReferencesReviewsResources

MxML: Mixture of Meta-Learners for Few-Shot Classification

Minseop Park, Jungtaek Kim, Saehoon Kim, Yanbin Liu, Seungjin Choi

Published 2019-04-11Version 1

A meta-model is trained on a distribution of similar tasks such that it learns an algorithm that can quickly adapt to a novel task with only a handful of labeled examples. Most of current meta-learning methods assume that the meta-training set consists of relevant tasks sampled from a single distribution. In practice, however, a new task is often out of the task distribution, yielding a performance degradation. One way to tackle this problem is to construct an ensemble of meta-learners such that each meta-learner is trained on different task distribution. In this paper we present a method for constructing a mixture of meta-learners (MxML), where mixing parameters are determined by the weight prediction network (WPN) optimized to improve the few-shot classification performance. Experiments on various datasets demonstrate that MxML significantly outperforms state-of-the-art meta-learners, or their naive ensemble in the case of out-of-distribution as well as in-distribution tasks.

Related articles: Most relevant | Search more
arXiv:2006.08875 [cs.LG] (Published 2020-06-16)
Model-based Adversarial Meta-Reinforcement Learning
arXiv:2206.10716 [cs.LG] (Published 2022-06-21)
Meta Reinforcement Learning with Finite Training Tasks -- a Density Estimation Approach
arXiv:2301.08028 [cs.LG] (Published 2023-01-19)
A Survey of Meta-Reinforcement Learning