Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Discriminative Sample-Guided and Parameter-Efficient Feature Space Adaptation for Cross-Domain Few-Shot Learning (2403.04492v3)

Published 7 Mar 2024 in cs.CV

Abstract: In this paper, we look at cross-domain few-shot classification which presents the challenging task of learning new classes in previously unseen domains with few labelled examples. Existing methods, though somewhat effective, encounter several limitations, which we alleviate through two significant improvements. First, we introduce a lightweight parameter-efficient adaptation strategy to address overfitting associated with fine-tuning a large number of parameters on small datasets. This strategy employs a linear transformation of pre-trained features, significantly reducing the trainable parameter count. Second, we replace the traditional nearest centroid classifier with a discriminative sample-aware loss function, enhancing the model's sensitivity to the inter- and intra-class variances within the training set for improved clustering in feature space. Empirical evaluations on the Meta-Dataset benchmark showcase that our approach not only improves accuracy up to 7.7\% and 5.3\% on previously seen and unseen datasets, respectively, but also achieves the above performance while being at least $\sim3\times$ more parameter-efficient than existing methods, establishing a new state-of-the-art in cross-domain few-shot learning. Our code is available at https://github.com/rashindrie/DIPA.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (40)
  1. Strong baselines for parameter efficient few-shot fine-tuning. CoRR, abs/2304.01917, 2023.
  2. Beyond simple meta-learning: Multi-purpose models for multi-domain, active and continual few-shot learning. CoRR, abs/2201.05151, 2022.
  3. Universal representations: The missing link between faces, text, planktons, and cat breeds. CoRR, abs/1701.07275, 2017a.
  4. Universal representations: The missing link between faces, text, planktons, and cat breeds. arXiv preprint arXiv:1701.07275, 2017b.
  5. Emerging properties in self-supervised vision transformers. In Proceedings of the IEEE/CVF international conference on computer vision, pages 9650–9660, 2021.
  6. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
  7. CrossTransformers: Spatially-aware few-shot transfer. Advances in Neural Information Processing Systems, 2020-Decem(NeurIPS), 2020a.
  8. Crosstransformers: spatially-aware few-shot transfer. Advances in Neural Information Processing Systems, 33:21981–21993, 2020b.
  9. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020.
  10. Timothy Dozat. Incorporating nesterov momentum into adam. 2016.
  11. Selecting Relevant Features from a Multi-domain Representation for Few-shot Classification. Eccv 2020, 12355 LNCS:769–786, 2020.
  12. One-shot learning of object categories. IEEE Trans. Pattern Anal. Mach. Intell., 28(4):594–611, 2006.
  13. Michael Fink. Object classification from a single example utilizing class relevance metrics. In Advances in Neural Information Processing Systems 17 [Neural Information Processing Systems, NIPS 2004, December 13-18, 2004, Vancouver, British Columbia, Canada], pages 449–456, 2004.
  14. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 16000–16009, 2022.
  15. Rethinking Generalization in Few-Shot Classification. (NeurIPS), 2022.
  16. Pushing the Limits of Simple Pipelines for Few-Shot Learning: External Data and Fine-Tuning Make a Difference. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2022-June:9058–9067, 2022.
  17. Proxy anchor loss for deep metric learning. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pages 3235–3244, 2020.
  18. Human-level concept learning through probabilistic program induction. Science, 350(6266):1332–1338, 2015.
  19. Cross-domain few-shot learning with task-specific adapters. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, June 18-24, 2022, pages 7151–7160. IEEE, 2022.
  20. Universal representation learning from multiple domains for few-shot classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 9526–9535, 2021.
  21. Deep metric learning for few-shot image classification: A review of recent developments. Pattern Recognition, page 109381, 2023.
  22. Scaling & Shifting Your Features: A New Baseline for Efficient Model Tuning. (NeurIPS):1–20, 2022.
  23. A universal representation transformer layer for few-shot image classification. 2021.
  24. Distance-based image classification: Generalizing to new classes at near-zero cost. IEEE transactions on pattern analysis and machine intelligence, 35(11):2624–2637, 2013.
  25. FiLM: Visual reasoning with a general conditioning layer. 32nd AAAI Conference on Artificial Intelligence, AAAI 2018, pages 3942–3951, 2018.
  26. Bi-level meta-learning for few-shot domain generalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 15900–15910, 2023.
  27. Efficient parametrization of multi-domain deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 8119–8127, 2018.
  28. Imagenet large scale visual recognition challenge. International journal of computer vision, 115:211–252, 2015.
  29. Prototypical networks for few-shot learning. Advances in neural information processing systems, 30, 2017.
  30. Similar class style augmentation for efficient cross-domain few-shot learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4589–4597, 2023.
  31. Return of frustratingly easy domain adaptation. In Proceedings of the AAAI conference on artificial intelligence, 2016.
  32. Meta-dataset: A dataset of datasets for learning to learn from few examples. 2020.
  33. Learning a universal template for few-shot dataset generalization. 139:10424–10433, 2021.
  34. Matching networks for one shot learning. In Advances in Neural Information Processing Systems, number Nips, pages 3637–3645, 2016.
  35. Exploring Efficient Few-shot Adaptation for Vision Transformers. (62076067), 2023.
  36. How transferable are features in deep neural networks? Advances in Neural Information Processing Systems, 4(January):3320–3328, 2014.
  37. Deepemd: Few-shot image classification with differentiable earth mover’s distance and structured classifiers. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 12203–12213, 2020.
  38. ibot: Image bert pre-training with online tokenizer. International Conference on Learning Representations (ICLR), 2022.
  39. Domain generalization with mixstyle. arXiv preprint arXiv:2104.02008, 2021.
  40. Revisiting mid-level patterns for cross-domain few-shot recognition. In Proceedings of the 29th ACM International Conference on Multimedia, pages 741–749, 2021.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Rashindrie Perera (2 papers)
  2. Saman Halgamuge (24 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.