Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Enhancing Out-of-Distribution Detection in Natural Language Understanding via Implicit Layer Ensemble (2210.11034v1)

Published 20 Oct 2022 in cs.CL and cs.LG

Abstract: Out-of-distribution (OOD) detection aims to discern outliers from the intended data distribution, which is crucial to maintaining high reliability and a good user experience. Most recent studies in OOD detection utilize the information from a single representation that resides in the penultimate layer to determine whether the input is anomalous or not. Although such a method is straightforward, the potential of diverse information in the intermediate layers is overlooked. In this paper, we propose a novel framework based on contrastive learning that encourages intermediate features to learn layer-specialized representations and assembles them implicitly into a single representation to absorb rich information in the pre-trained LLM. Extensive experiments in various intent classification and OOD datasets demonstrate that our approach is significantly more effective than other works.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Hyunsoo Cho (28 papers)
  2. Choonghyun Park (6 papers)
  3. Jaewook Kang (15 papers)
  4. Kang Min Yoo (40 papers)
  5. Taeuk Kim (38 papers)
  6. Sang-goo Lee (40 papers)
Citations (8)