FREE: Faster and Better Data-Free Meta-Learning (2405.00984v2)
Abstract: Data-Free Meta-Learning (DFML) aims to extract knowledge from a collection of pre-trained models without requiring the original data, presenting practical benefits in contexts constrained by data privacy concerns. Current DFML methods primarily focus on the data recovery from these pre-trained models. However, they suffer from slow recovery speed and overlook gaps inherent in heterogeneous pre-trained models. In response to these challenges, we introduce the Faster and Better Data-Free Meta-Learning (FREE) framework, which contains: (i) a meta-generator for rapidly recovering training tasks from pre-trained models; and (ii) a meta-learner for generalizing to new unseen tasks. Specifically, within the module Faster Inversion via Meta-Generator, each pre-trained model is perceived as a distinct task. The meta-generator can rapidly adapt to a specific task in just five steps, significantly accelerating the data recovery. Furthermore, we propose Better Generalization via Meta-Learner and introduce an implicit gradient alignment algorithm to optimize the meta-learner. This is achieved as aligned gradient directions alleviate potential conflicts among tasks from heterogeneous pre-trained models. Empirical experiments on multiple benchmarks affirm the superiority of our approach, marking a notable speed-up (20$\times$) and performance enhancement (1.42%$\sim$4.78%) in comparison to the state-of-the-art.
- Meta-learning with task-adaptive loss function for few-shot learning. In ICCV, pages 9465–9474, 2021.
- Meta-learning with differentiable closed-form solvers. In ICLR, 2019.
- Dynamic kernel selection for improved generalization and memory efficiency in meta-learning. In CVPR, pages 9851–9860, 2022.
- Variational attention: Propagating domain-specific knowledge for multi-domain learning in crowd counting. In ICCV, pages 16065–16075, 2021a.
- Data-free learning of student networks. In ICCV, pages 3514–3522, 2019.
- Variational metric scaling for metric-based meta-learning. In AAAI, pages 3478–3485, 2020.
- Meta-baseline: Exploring simple meta-learning for few-shot learning. In ICCV, pages 9062–9071, 2021b.
- Implicit gradient alignment in distributed and federated learning. In AAAI, pages 6454–6462, 2022.
- Gradient agreement as an optimization objective for meta-learning. arXiv preprint arXiv:1810.08178, 2018.
- Up to 100x faster data-free knowledge distillation. In AAAI, pages 6597–6604, 2022.
- Model-agnostic meta-learning for fast adaptation of deep networks. In ICML, pages 1126–1135, 2017.
- Towards data-free domain generalization. In ACML, pages 327–342, 2023.
- Imagenet-trained cnns are biased towards texture; increasing shape bias improves accuracy and robustness. In ICLR, 2018.
- Shortcut learning in deep neural networks. Nature Machine Intelligence, 2:665–673, 2020.
- Memory-efficient backpropagation through time. In NeurIPS, 2016.
- Recon: Reducing conflicting gradients from the root for multi-task learning. In ICLR, 2022.
- One-for-all: Bridge the gap between heterogeneous architectures in knowledge distillation. In NeurIPS, 2023.
- Architecture, dataset and model-scale agnostic data-free meta-learning. In CVPR, pages 7736–7745, 2023a.
- Zixuan Hu et al. Learning to learn from apis:black-box data-free meta-learning. In ICML, 2023b.
- Task agnostic meta-learning for few-shot learning. In CVPR, pages 11719–11727, 2019.
- Similarity of neural network representations revisited. In ICML, pages 3519–3529, 2019.
- Repurposing pretrained models for robust out-of-domain few-shot learning. In ICLR, 2020.
- Contextual gradient scaling for few-shot learning. In WACV, pages 834–843, 2022.
- Deep model fusion: A survey. arXiv preprint arXiv:2309.15698, 2023.
- Adaptive task sampling for meta-learning. In ECCV, pages 752–769, 2020.
- Data-free knowledge transfer: A survey. arXiv preprint arXiv:2112.15278, 2021.
- Zero-shot knowledge transfer via adversarial belief matching. In NeurIPS, 2019.
- On first-order meta-learning algorithms. arXiv preprint arXiv:1803.02999, 2018.
- Learning to retain while acquiring: Combating distribution-shift in adversarial data-free knowledge distillation. In CVPR, pages 7786–7794, 2023.
- Bi-level meta-learning for few-shot domain generalization. In CVPR, pages 15900–15910, 2023.
- Rapid learning or feature reuse? towards understanding the effectiveness of maml. In ICLR, 2019.
- itaml: An incremental task-agnostic meta-learning approach. In CVPR, pages 13588–13597, 2020.
- Meta-learning with implicit gradients. In NeurIPS, 2019.
- Optimization as a model for few-shot learning. In ICLR, 2017.
- Learning to reweight examples for robust deep learning. In ICML, pages 4334–4343, 2018.
- Towards data-free model stealing in a hard label setting. In CVPR, pages 15284–15293, 2022.
- Gradient matching for domain generalization. In ICLR, 2021.
- Model fusion via optimal transport. In NeurIPS, pages 22045–22055, 2020.
- Prototypical networks for few-shot learning. In NeurIPS, pages 4077–4087, 2017.
- Robust texture description using local grouped order pattern and non-local binary pattern. IEEE TCSVT, 31:189–202, 2020.
- Data-free model extraction. In CVPR, pages 4771–4780, 2021.
- The Caltech-UCSD birds-200-2011 dataset. Technical report, California Institute of Technology, 2011.
- Learning to learn and remember super long multi-domain task sequence. In CVPR, pages 7982–7992, 2022a.
- Meta-learning without data via wasserstein distributionally-robust model fusion. In UAI, pages 2045–2055, 2022b.
- Few-shot classification with feature map reconstruction networks. In CVPR, pages 8012–8021, 2021.
- Hierarchically structured meta-learning. In ICML, pages 7045–7054, 2019.
- Online structured meta-learning. In NeurIPS, 2020.
- Meta-learning with an adaptive task scheduler. In NeurIPS, pages 7497–7509, 2021.
- Dreaming to distill: Data-free knowledge transfer via deepinversion. In CVPR, pages 8715–8724, 2020.
- Data-free knowledge distillation via feature exchange and activation region constraint. In CVPR, pages 24266–24275, 2023.
- Gradient surgery for multi-task learning. In NeurIPS, pages 5824–5836, 2020.
- Learn from model beyond fine-tuning: A survey. arXiv preprint arXiv:2310.08184, 2023.