Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

From Audio to Semantics: Approaches to end-to-end spoken language understanding (1809.09190v1)

Published 24 Sep 2018 in eess.AS, cs.CL, and cs.SD

Abstract: Conventional spoken language understanding systems consist of two main components: an automatic speech recognition module that converts audio to a transcript, and a natural language understanding module that transforms the resulting text (or top N hypotheses) into a set of domains, intents, and arguments. These modules are typically optimized independently. In this paper, we formulate audio to semantic understanding as a sequence-to-sequence problem [1]. We propose and compare various encoder-decoder based approaches that optimize both modules jointly, in an end-to-end manner. Evaluations on a real-world task show that 1) having an intermediate text representation is crucial for the quality of the predicted semantics, especially the intent arguments and 2) jointly optimizing the full system improves overall accuracy of prediction. Compared to independently trained models, our best jointly trained model achieves similar domain and intent prediction F1 scores, but improves argument word error rate by 18% relative.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Parisa Haghani (15 papers)
  2. Arun Narayanan (34 papers)
  3. Michiel Bacchiani (16 papers)
  4. Galen Chuang (3 papers)
  5. Neeraj Gaur (7 papers)
  6. Pedro Moreno (10 papers)
  7. Rohit Prabhavalkar (59 papers)
  8. Zhongdi Qu (5 papers)
  9. Austin Waters (10 papers)
Citations (148)

Summary

We haven't generated a summary for this paper yet.