Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Improving Spoken Language Understanding By Exploiting ASR N-best Hypotheses (2001.05284v1)

Published 11 Jan 2020 in cs.CL, cs.SD, and eess.AS

Abstract: In a modern spoken language understanding (SLU) system, the natural language understanding (NLU) module takes interpretations of a speech from the automatic speech recognition (ASR) module as the input. The NLU module usually uses the first best interpretation of a given speech in downstream tasks such as domain and intent classification. However, the ASR module might misrecognize some speeches and the first best interpretation could be erroneous and noisy. Solely relying on the first best interpretation could make the performance of downstream tasks non-optimal. To address this issue, we introduce a series of simple yet efficient models for improving the understanding of semantics of the input speeches by collectively exploiting the n-best speech interpretations from the ASR module.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Mingda Li (95 papers)
  2. Weitong Ruan (3 papers)
  3. Xinyue Liu (24 papers)
  4. Luca Soldaini (62 papers)
  5. Wael Hamza (26 papers)
  6. Chengwei Su (5 papers)
Citations (22)