How to Train Your MAML to Excel in Few-Shot Classification

06/30/2021
by   Han-Jia Ye, et al.
0

Model-agnostic meta-learning (MAML) is arguably the most popular meta-learning algorithm nowadays, given its flexibility to incorporate various model architectures and to be applied to different problems. Nevertheless, its performance on few-shot classification is far behind many recent algorithms dedicated to the problem. In this paper, we point out several key facets of how to train MAML to excel in few-shot classification. First, we find that a large number of gradient steps are needed for the inner loop update, which contradicts the common usage of MAML for few-shot classification. Second, we find that MAML is sensitive to the permutation of class assignments in meta-testing: for a few-shot task of N classes, there are exponentially many ways to assign the learned initialization of the N-way classifier to the N classes, leading to an unavoidably huge variance. Third, we investigate several ways for permutation invariance and find that learning a shared classifier initialization for all the classes performs the best. On benchmark datasets such as MiniImageNet and TieredImageNet, our approach, which we name UNICORN-MAML, performs on a par with or even outperforms state-of-the-art algorithms, while keeping the simplicity of MAML without adding any extra sub-networks.

READ FULL TEXT
research
11/28/2018

Unsupervised Meta-Learning For Few-Shot Image and Video Classification

Few-shot or one-shot learning of classifiers for images or videos is an ...
research
06/28/2020

Many-Class Few-Shot Learning on Multi-Granularity Class Hierarchy

We study many-class few-shot (MCFS) problem in both supervised learning ...
research
09/20/2022

MAC: A Meta-Learning Approach for Feature Learning and Recombination

Optimization-based meta-learning aims to learn an initialization so that...
research
10/20/2021

Contextual Gradient Scaling for Few-Shot Learning

Model-agnostic meta-learning (MAML) is a well-known optimization-based m...
research
02/14/2021

Large-Scale Meta-Learning with Continual Trajectory Shifting

Meta-learning of shared initialization parameters has shown to be highly...
research
02/08/2019

Adversarial Initialization -- when your network performs the way I want

The increase in computational power and available data has fueled a wide...
research
05/22/2023

Mitigating Catastrophic Forgetting for Few-Shot Spoken Word Classification Through Meta-Learning

We consider the problem of few-shot spoken word classification in a sett...

Please sign up or login with your details

Forgot password? Click here to reset