Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Towards Robust and Generalizable Training: An Empirical Study of Noisy Slot Filling for Input Perturbations (2310.03518v1)

Published 5 Oct 2023 in cs.CL, cs.AI, and cs.DS

Abstract: In real dialogue scenarios, as there are unknown input noises in the utterances, existing supervised slot filling models often perform poorly in practical applications. Even though there are some studies on noise-robust models, these works are only evaluated on rule-based synthetic datasets, which is limiting, making it difficult to promote the research of noise-robust methods. In this paper, we introduce a noise robustness evaluation dataset named Noise-SF for slot filling task. The proposed dataset contains five types of human-annotated noise, and all those noises are exactly existed in real extensive robust-training methods of slot filling into the proposed framework. By conducting exhaustive empirical evaluation experiments on Noise-SF, we find that baseline models have poor performance in robustness evaluation, and the proposed framework can effectively improve the robustness of models. Based on the empirical experimental results, we make some forward-looking suggestions to fuel the research in this direction. Our dataset Noise-SF will be released at https://github.com/dongguanting/Noise-SF.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. Jiachi Liu (5 papers)
  2. Liwen Wang (18 papers)
  3. Guanting Dong (46 papers)
  4. Xiaoshuai Song (16 papers)
  5. Zechen Wang (15 papers)
  6. Zhengyang Wang (48 papers)
  7. Shanglin Lei (5 papers)
  8. Jinzheng Zhao (18 papers)
  9. Keqing He (47 papers)
  10. Bo Xiao (62 papers)
  11. Weiran Xu (58 papers)
Citations (6)
Github Logo Streamline Icon: https://streamlinehq.com