Meta-learning with attention for improved few-shot learning

Zejiang Hou, Anwar Walid, Sun Yuan Kung

Research output: Contribution to journalConference articlepeer-review

Abstract

We consider few-shot learning (FSL), where a model learns from very few labeled examples such that it can generalize to unseen examples. Model-agnostic meta-learning (MAML) has been proposed to solve FSL. However, the low performance of MAML suggests its difficulty in tackle diverse tasks, due to the restriction of sharing a single model initialization for fast adaptation. In this paper, we propose meta-learning with attention mechanisms. Our method meta-learns attention modules to instantiate task-specific model initialization for fast adaptation, which can obtain high-quality solution to a new task using few gradient descent steps. To further improve generalization during inference, we propose to incorporate an entropy regularizer into the adaptation objective to penalize the Shannon entropy of prediction probability. Extensive experiments under various FSL scenarios show that our method achieves state-of-the-art performance on the mini-ImageNet and tiered-ImageNet.

Original languageEnglish (US)
Pages (from-to)2725-2729
Number of pages5
JournalICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
Volume2021-June
DOIs
StatePublished - 2021
Event2021 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2021 - Virtual, Toronto, Canada
Duration: Jun 6 2021Jun 11 2021

All Science Journal Classification (ASJC) codes

  • Software
  • Signal Processing
  • Electrical and Electronic Engineering

Keywords

  • Attention mechanism
  • Few-shot learning
  • Meta-learning

Fingerprint

Dive into the research topics of 'Meta-learning with attention for improved few-shot learning'. Together they form a unique fingerprint.

Cite this