Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Repurposing Pretrained Models for Robust Out-of-domain Few-Shot Learning (2103.09027v1)

Published 16 Mar 2021 in cs.LG and cs.CV

Abstract: Model-agnostic meta-learning (MAML) is a popular method for few-shot learning but assumes that we have access to the meta-training set. In practice, training on the meta-training set may not always be an option due to data privacy concerns, intellectual property issues, or merely lack of computing resources. In this paper, we consider the novel problem of repurposing pretrained MAML checkpoints to solve new few-shot classification tasks. Because of the potential distribution mismatch, the original MAML steps may no longer be optimal. Therefore we propose an alternative meta-testing procedure and combine MAML gradient steps with adversarial training and uncertainty-based stepsize adaptation. Our method outperforms "vanilla" MAML on same-domain and cross-domains benchmarks using both SGD and Adam optimizers and shows improved robustness to the choice of base stepsize.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Namyeong Kwon (1 paper)
  2. Hwidong Na (4 papers)
  3. Gabriel Huang (10 papers)
  4. Simon Lacoste-Julien (95 papers)
Citations (7)

Summary

We haven't generated a summary for this paper yet.