Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

From Disfluency Detection to Intent Detection and Slot Filling (2209.08359v1)

Published 17 Sep 2022 in cs.CL

Abstract: We present the first empirical study investigating the influence of disfluency detection on downstream tasks of intent detection and slot filling. We perform this study for Vietnamese -- a low-resource language that has no previous study as well as no public dataset available for disfluency detection. First, we extend the fluent Vietnamese intent detection and slot filling dataset PhoATIS by manually adding contextual disfluencies and annotating them. Then, we conduct experiments using strong baselines for disfluency detection and joint intent detection and slot filling, which are based on pre-trained LLMs. We find that: (i) disfluencies produce negative effects on the performances of the downstream intent detection and slot filling tasks, and (ii) in the disfluency context, the pre-trained multilingual LLM XLM-R helps produce better intent detection and slot filling performances than the pre-trained monolingual LLM PhoBERT, and this is opposite to what generally found in the fluency context.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Mai Hoang Dao (5 papers)
  2. Thinh Hung Truong (9 papers)
  3. Dat Quoc Nguyen (55 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.