Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Empirical Bayes Transductive Meta-Learning with Synthetic Gradients (2004.12696v1)

Published 27 Apr 2020 in cs.LG and stat.ML

Abstract: We propose a meta-learning approach that learns from multiple tasks in a transductive setting, by leveraging the unlabeled query set in addition to the support set to generate a more powerful model for each task. To develop our framework, we revisit the empirical Bayes formulation for multi-task learning. The evidence lower bound of the marginal log-likelihood of empirical Bayes decomposes as a sum of local KL divergences between the variational posterior and the true posterior on the query set of each task. We derive a novel amortized variational inference that couples all the variational posteriors via a meta-model, which consists of a synthetic gradient network and an initialization network. Each variational posterior is derived from synthetic gradient descent to approximate the true posterior on the query set, although where we do not have access to the true gradient. Our results on the Mini-ImageNet and CIFAR-FS benchmarks for episodic few-shot classification outperform previous state-of-the-art methods. Besides, we conduct two zero-shot learning experiments to further explore the potential of the synthetic gradient.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Shell Xu Hu (18 papers)
  2. Pablo G. Moreno (5 papers)
  3. Yang Xiao (149 papers)
  4. Xi Shen (46 papers)
  5. Guillaume Obozinski (27 papers)
  6. Neil D. Lawrence (68 papers)
  7. Andreas Damianou (28 papers)
Citations (119)