Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Towards Few-shot Out-of-Distribution Detection (2311.12076v3)

Published 20 Nov 2023 in cs.CV

Abstract: Out-of-distribution (OOD) detection is critical for ensuring the reliability of open-world intelligent systems. Despite the notable advancements in existing OOD detection methodologies, our study identifies a significant performance drop under the scarcity of training samples. In this context, we introduce a novel few-shot OOD detection benchmark, carefully constructed to address this gap. Our empirical analysis reveals the superiority of ParameterEfficient Fine-Tuning (PEFT) strategies, such as visual prompt tuning and visual adapter tuning, over conventional techniques, including fully fine-tuning and linear probing tuning in the few-shot OOD detection task. Recognizing some crucial information from the pre-trained model, which is pivotal for OOD detection, may be lost during the fine-tuning process, we propose a method termed DomainSpecific and General Knowledge Fusion (DSGF). This approach is designed to be compatible with diverse fine-tuning frameworks. Our experiments show that the integration of DSGF significantly enhances the few-shot OOD detection capabilities across various methods and fine-tuning methodologies, including fully fine-tuning, visual adapter tuning, and visual prompt tuning. The code will be released.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (47)
  1. Concrete problems in ai safety. arXiv preprint arXiv:1606.06565, 2016.
  2. Food-101–mining discriminative components with random forests. In Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part VI 13, pp.  446–461. Springer, 2014.
  3. Emerging properties in self-supervised vision transformers. In Proceedings of the IEEE/CVF international conference on computer vision, pp.  9650–9660, 2021.
  4. Describing textures in the wild. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp.  3606–3613, 2014.
  5. Leveraging visual attention for out-of-distribution detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp.  4447–4456, 2023.
  6. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pp.  248–255. Ieee, 2009.
  7. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020.
  8. Vos: Learning what you don’t know by virtual outlier synthesis. In International Conference on Learning Representations, 2021.
  9. Zero-shot out-of-distribution detection based on the pre-trained model clip. In Proceedings of the AAAI conference on artificial intelligence, volume 36, pp.  6568–6576, 2022.
  10. On the effectiveness of parameter-efficient fine-tuning. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pp.  12799–12807, 2023.
  11. A survey of accuracy evaluation metrics of recommendation tasks. Journal of Machine Learning Research, 10(12), 2009.
  12. A baseline for detecting misclassified and out-of-distribution examples in neural networks. In International Conference on Learning Representations, 2016.
  13. Scaling out-of-distribution detection for real-world settings. In International Conference on Machine Learning, pp.  8759–8773. PMLR, 2022.
  14. Out-of-distribution detection as support for autonomous driving safety lifecycle. In International Working Conference on Requirements Engineering: Foundation for Software Quality, pp.  233–242. Springer, 2023.
  15. Parameter-efficient transfer learning for nlp. In International Conference on Machine Learning, pp.  2790–2799. PMLR, 2019.
  16. Mos: Towards scaling out-of-distribution detection for large semantic space. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  8710–8719, 2021.
  17. Visual prompt tuning. In European Conference on Computer Vision, pp.  709–727. Springer, 2022.
  18. Rethinking efficient tuning methods from a unified perspective. arXiv preprint arXiv:2303.00690, 2023.
  19. Segment anything. arXiv preprint arXiv:2304.02643, 2023.
  20. Do better imagenet models transfer better? In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp.  2661–2671, 2019.
  21. Krizhevsky, A. et al. Learning multiple layers of features from tiny images. 2009.
  22. Tiny imagenet visual recognition challenge. CS 231N, 7(7):3, 2015.
  23. The power of scale for parameter-efficient prompt tuning. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp.  3045–3059, 2021.
  24. Enhancing the reliability of out-of-distribution image detection in neural networks. In International Conference on Learning Representations, 2018.
  25. Mood: Multi-level out-of-distribution detection. In Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition, pp.  15313–15323, 2021.
  26. Transfer learning with shapeshift adapter: A parameter-efficient module for deep learning model. In 2020 International Conference on Machine Learning and Cybernetics (ICMLC), pp.  105–110. IEEE, 2020.
  27. Energy-based out-of-distribution detection. Advances in neural information processing systems, 33:21464–21475, 2020.
  28. Few-shot keypoint detection with uncertainty learning for unseen species. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  19416–19426, 2022.
  29. Delving into out-of-distribution detection with vision-language representations. Advances in Neural Information Processing Systems, 35:35087–35102, 2022.
  30. Locoop: Few-shot out-of-distribution detection via prompt learning. In Thirty-seventh Conference on Neural Information Processing Systems, 2023. URL https://openreview.net/forum?id=UjtiLdXGMC.
  31. Dinov2: Learning robust visual features without supervision. arXiv preprint arXiv:2304.07193, 2023.
  32. Cats and dogs. In 2012 IEEE conference on computer vision and pattern recognition, pp.  3498–3505. IEEE, 2012.
  33. Adapterhub: A framework for adapting transformers. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pp.  46–54, 2020.
  34. Powers, D. Evaluation: From precision, recall and f-measure to roc, informedness, markedness & correlation. Journal of Machine Learning Technologies, 2(1):37–63, 2011.
  35. Learning transferable visual models from natural language supervision. In International conference on machine learning, pp.  8748–8763. PMLR, 2021.
  36. Likelihood ratios for out-of-distribution detection. Advances in neural information processing systems, 32, 2019.
  37. Out-of-domain detection based on generative adversarial network. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp.  714–718, 2018.
  38. Out-of-distribution detection with deep nearest neighbors. In International Conference on Machine Learning, pp.  20827–20840. PMLR, 2022.
  39. Visualizing data using t-sne. Journal of machine learning research, 9(11), 2008.
  40. The inaturalist species classification and detection dataset. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp.  8769–8778, 2018.
  41. Sun database: Large-scale scene recognition from abbey to zoo. In 2010 IEEE computer society conference on computer vision and pattern recognition, pp.  3485–3492. IEEE, 2010.
  42. Openood: Benchmarking generalized out-of-distribution detection. In Thirty-sixth Conference on Neural Information Processing Systems Datasets and Benchmarks Track, 2022.
  43. Aim: Adapting image models for efficient video action recognition. In The Eleventh International Conference on Learning Representations, 2023.
  44. Glipv2: Unifying localization and vision-language understanding. In Advances in Neural Information Processing Systems, 2022.
  45. Places: A 10 million image database for scene recognition. IEEE transactions on pattern analysis and machine intelligence, 40(6):1452–1464, 2017.
  46. Conditional prompt learning for vision-language models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  16816–16825, 2022a.
  47. Learning to prompt for vision-language models. International Journal of Computer Vision, 130(9):2337–2348, 2022b.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Jiuqing Dong (1 paper)
  2. Yongbin Gao (6 papers)
  3. Heng Zhou (47 papers)
  4. Jun Cen (28 papers)
  5. Yifan Yao (11 papers)
  6. Sook Yoon (7 papers)
  7. Park Dong Sun (1 paper)
Citations (3)

Summary

We haven't generated a summary for this paper yet.