Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Mixture Outlier Exposure: Towards Out-of-Distribution Detection in Fine-grained Environments (2106.03917v4)

Published 7 Jun 2021 in cs.LG

Abstract: Many real-world scenarios in which DNN-based recognition systems are deployed have inherently fine-grained attributes (e.g., bird-species recognition, medical image classification). In addition to achieving reliable accuracy, a critical subtask for these models is to detect Out-of-distribution (OOD) inputs. Given the nature of the deployment environment, one may expect such OOD inputs to also be fine-grained w.r.t. the known classes (e.g., a novel bird species), which are thus extremely difficult to identify. Unfortunately, OOD detection in fine-grained scenarios remains largely underexplored. In this work, we aim to fill this gap by first carefully constructing four large-scale fine-grained test environments, in which existing methods are shown to have difficulties. Particularly, we find that even explicitly incorporating a diverse set of auxiliary outlier data during training does not provide sufficient coverage over the broad region where fine-grained OOD samples locate. We then propose Mixture Outlier Exposure (MixOE), which mixes ID data and training outliers to expand the coverage of different OOD granularities, and trains the model such that the prediction confidence linearly decays as the input transitions from ID to OOD. Extensive experiments and analyses demonstrate the effectiveness of MixOE for building up OOD detector in fine-grained environments. The code is available at https://github.com/zjysteven/MixOE.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Jingyang Zhang (58 papers)
  2. Nathan Inkawhich (19 papers)
  3. Randolph Linderman (5 papers)
  4. Yiran Chen (176 papers)
  5. Hai Li (159 papers)
Citations (36)

Summary

The paper "Mixture Outlier Exposure: Towards Out-of-Distribution Detection in Fine-grained Environments" addresses the challenge of out-of-distribution (OOD) detection in fine-grained contexts, such as bird species recognition or medical image classification, where the differences between categories are subtle and often visually similar. Despite the presence of deep neural network (DNN) models designed for high accuracy in classification, detecting fine-grained OOD inputs—like novel species not included in training—remains difficult due to their semantic closeness to known categories.

The authors identify the lack of research in OOD detection in such fine-grained environments and propose a new method called Mixture Outlier Exposure (MixOE). This approach aims to improve the model's ability to detect OOD inputs by creating a range of virtual outlier samples that span different levels of similarity to in-distribution (ID) data, thereby enhancing the OOD detection capability across a spectrum of granularities.

Key points from the paper include:

  1. Construction of Test Environments: The authors construct four large-scale fine-grained test environments using publicly available datasets such as FGVC-Aircraft, Stanford Cars, Butterfly, and North American Birds. They create ID and OOD splits that better approximate real-world scenarios for evaluating OOD detection methods.
  2. Challenges in Fine-grained OOD Detection: Initial evaluations show that existing state-of-the-art OOD detection methods struggle significantly with fine-grained OOD data, performing well on coarse-grained but not on semantically similar inputs. Traditional methods like MSP, ODIN, and Energy, and even those incorporating training with outlier data, do not reliably improve performance against fine-grained novelties.
  3. Mixture Outlier Exposure (MixOE): MixOE utilizes both ID data and auxiliary outlier data during training to form virtual outlier samples through a mixing process (such as Mixup or CutMix operations). This process is aimed at expanding the coverage of possible OOD inputs in feature space, thus enhancing detection across a gradient of input similarities.
  4. Training Strategy and Implication: The method trains the model to progressively decay its prediction confidence from ID to OOD samples, thus covering both coarse- and fine-grained inputs. MixOE employs unlabeled auxiliary datasets like WebVision to create diverse virtual outlier scenarios without incorporating concepts related to actual test OOD distributions.
  5. Experiment and Validation: MixOE demonstrates notable improvements in TNR95 (True Negative Rate at 95% True Positive Rate) compared to other methods across all datasets, achieving better or comparable detection rates for both coarse- and fine-grained OOD inputs without sacrificing in-distribution classification accuracy. The method uniquely balances detection effectiveness across different granular levels where other methods fall short.
  6. Conclusion: MixOE provides a scalable solution to the OOD detection problem in finely detailed recognition tasks. It significantly enhances the models' capability to recognize unseen categories that are subtly different from those they were trained on, thereby making intelligent systems safer and more robust in open-world applications.

The paper highlights MixOE as a promising technique for dealing with fine-grained OOD detection, encouraging future research to further explore and develop methodologies for such challenging recognition tasks.