Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Detecting Out-of-Distribution Samples via Conditional Distribution Entropy with Optimal Transport (2401.11726v1)

Published 22 Jan 2024 in cs.LG and cs.CV

Abstract: When deploying a trained machine learning model in the real world, it is inevitable to receive inputs from out-of-distribution (OOD) sources. For instance, in continual learning settings, it is common to encounter OOD samples due to the non-stationarity of a domain. More generally, when we have access to a set of test inputs, the existing rich line of OOD detection solutions, especially the recent promise of distance-based methods, falls short in effectively utilizing the distribution information from training samples and test inputs. In this paper, we argue that empirical probability distributions that incorporate geometric information from both training samples and test inputs can be highly beneficial for OOD detection in the presence of test inputs available. To address this, we propose to model OOD detection as a discrete optimal transport problem. Within the framework of optimal transport, we propose a novel score function known as the \emph{conditional distribution entropy} to quantify the uncertainty of a test input being an OOD sample. Our proposal inherits the merits of certain distance-based methods while eliminating the reliance on distribution assumptions, a-prior knowledge, and specific training mechanisms. Extensive experiments conducted on benchmark datasets demonstrate that our method outperforms its competitors in OOD detection.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (43)
  1. Introduction to linear optimization, volume 6. Athena Scientific Belmont, MA, 1997.
  2. Convex optimization. Cambridge university press, 2004.
  3. A simple framework for contrastive learning of visual representations. In International conference on machine learning, pp.  1597–1607. PMLR, 2020.
  4. Describing textures in the wild. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp.  3606–3613, 2014.
  5. Marco Cuturi. Sinkhorn distances: Lightspeed computation of optimal transport. In NIPS, 2013.
  6. Is out-of-distribution detection learnable? ArXiv, abs/2210.14707, 2022.
  7. Exploring the limits of out-of-distribution detection. Advances in Neural Information Processing Systems, 34:7068–7081, 2021.
  8. In-or out-of-distribution detection via dual divergence estimation. In Uncertainty in Artificial Intelligence, pp.  635–646. PMLR, 2023.
  9. Unsupervised representation learning by predicting image rotations. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=S1v4N2l0-.
  10. Outlier detection using replicator neural networks. In Data Warehousing and Knowledge Discovery: 4th International Conference, DaWaK 2002 Aix-en-Provence, France, September 4–6, 2002 Proceedings 4, pp.  170–180. Springer, 2002.
  11. Identity mappings in deep residual networks. In European conference on computer vision, pp.  630–645. Springer, 2016.
  12. A baseline for detecting misclassified and out-of-distribution examples in neural networks. Proceedings of International Conference on Learning Representations, 2017.
  13. Deep anomaly detection with outlier exposure. In International Conference on Learning Representations, 2019. URL https://openreview.net/forum?id=HyxCxhRcY7.
  14. Generalized odin: Detecting out-of-distribution image without learning from out-of-distribution data. 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.  10948–10957, 2020.
  15. Supervised contrastive learning. Advances in Neural Information Processing Systems, 33:18661–18673, 2020.
  16. Proxy anchor loss for deep metric learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  3238–3247, 2020.
  17. Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114, 2013.
  18. Auto-Encoding Variational Bayes. In ICLR, 2014.
  19. Learning multiple layers of features from tiny images. 2009.
  20. Brian Kulis et al. Metric learning: A survey. Foundations and Trends® in Machine Learning, 5(4):287–364, 2013.
  21. A simple unified framework for detecting out-of-distribution samples and adversarial attacks. Advances in neural information processing systems, 31, 2018.
  22. Enhancing the reliability of out-of-distribution image detection in neural networks. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=H1VGkIxRZ.
  23. Energy-based out-of-distribution detection. Advances in Neural Information Processing Systems, 33:21464–21475, 2020.
  24. Out-of-distribution detection in high-dimensional data using mahalanobis distance-critical analysis. In Computational Science–ICCS 2022: 22nd International Conference, London, UK, June 21–23, 2022, Proceedings, Part I, pp.  262–275. Springer, 2022.
  25. Poem: Out-of-distribution detection with posterior sampling. In International Conference on Machine Learning, 2022.
  26. How to exploit hyperspherical embeddings for out-of-distribution detection? In Proceedings of the International Conference on Learning Representations, 2023.
  27. Provable guarantees for understanding out-of-distribution detection. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pp.  7831–7840, 2022.
  28. Towards maximizing the representation gap between in-domain & out-of-distribution examples. Advances in Neural Information Processing Systems, 33:9239–9250, 2020.
  29. Reading digits in natural images with unsupervised feature learning. 2011.
  30. Computational optimal transport. Found. Trends Mach. Learn., 11:355–607, 2018.
  31. Deep one-class classification. In International Conference on Machine Learning, 2018.
  32. PixelCNN++: Improving the pixelCNN with discretized logistic mixture likelihood and other modifications. In International Conference on Learning Representations, 2017. URL https://openreview.net/forum?id=BJrFC6ceg.
  33. Detecting out-of-distribution examples with gram matrices. In International Conference on Machine Learning, pp.  8491–8501. PMLR, 2020.
  34. Ssd: A unified framework for self-supervised outlier detection. In International Conference on Learning Representations, 2021. URL https://openreview.net/forum?id=v5gjXpmR8J.
  35. Concerning nonnegative matrices and doubly stochastic matrices. Pacific Journal of Mathematics, 21(2):343–348, 1967.
  36. Out-of-distribution detection with deep nearest neighbors. In International Conference on Machine Learning, 2022.
  37. Csi: Novelty detection via contrastive learning on distributionally shifted instances. Advances in neural information processing systems, 33:11839–11852, 2020.
  38. Terence Tao. An introduction to measure theory, volume 126. American Mathematical Society Providence, 2011.
  39. Cédric Villani. Topics in optimal transportation. 2003.
  40. Distance metric learning for large margin nearest neighbor classification. Journal of machine learning research, 10(2), 2009.
  41. Contrastive training for improved out-of-distribution detection. ArXiv, abs/2007.05566, 2020.
  42. Lsun: Construction of a large-scale image dataset using deep learning with humans in the loop. ArXiv, abs/1506.03365, 2015.
  43. Places: A 10 million image database for scene recognition. IEEE transactions on pattern analysis and machine intelligence, 40(6):1452–1464, 2017.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Chuanwen Feng (2 papers)
  2. Wenlong Chen (15 papers)
  3. Ao Ke (3 papers)
  4. Yilong Ren (17 papers)
  5. Xike Xie (21 papers)
  6. S. Kevin Zhou (165 papers)

Summary

We haven't generated a summary for this paper yet.