Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Attentive Contextual Carryover for Multi-Turn End-to-End Spoken Language Understanding (2112.06743v1)

Published 13 Dec 2021 in cs.CL and cs.AI

Abstract: Recent years have seen significant advances in end-to-end (E2E) spoken language understanding (SLU) systems, which directly predict intents and slots from spoken audio. While dialogue history has been exploited to improve conventional text-based natural language understanding systems, current E2E SLU approaches have not yet incorporated such critical contextual signals in multi-turn and task-oriented dialogues. In this work, we propose a contextual E2E SLU model architecture that uses a multi-head attention mechanism over encoded previous utterances and dialogue acts (actions taken by the voice assistant) of a multi-turn dialogue. We detail alternative methods to integrate these contexts into the state-ofthe-art recurrent and transformer-based models. When applied to a large de-identified dataset of utterances collected by a voice assistant, our method reduces average word and semantic error rates by 10.8% and 12.6%, respectively. We also present results on a publicly available dataset and show that our method significantly improves performance over a noncontextual baseline

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. Kai Wei (30 papers)
  2. Thanh Tran (52 papers)
  3. Feng-Ju Chang (15 papers)
  4. Kanthashree Mysore Sathyendra (10 papers)
  5. Thejaswi Muniyappa (4 papers)
  6. Jing Liu (525 papers)
  7. Anirudh Raju (20 papers)
  8. Ross McGowan (4 papers)
  9. Nathan Susanj (12 papers)
  10. Ariya Rastrow (55 papers)
  11. Grant P. Strimel (21 papers)
Citations (10)