Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Rethinking Label Smoothing on Multi-hop Question Answering (2212.09512v3)

Published 19 Dec 2022 in cs.CL

Abstract: Multi-Hop Question Answering (MHQA) is a significant area in question answering, requiring multiple reasoning components, including document retrieval, supporting sentence prediction, and answer span extraction. In this work, we analyze the primary factors limiting the performance of multi-hop reasoning and introduce label smoothing into the MHQA task. This is aimed at enhancing the generalization capabilities of MHQA systems and mitigating overfitting of answer spans and reasoning paths in training set. We propose a novel label smoothing technique, F1 Smoothing, which incorporates uncertainty into the learning process and is specifically tailored for Machine Reading Comprehension (MRC) tasks. Inspired by the principles of curriculum learning, we introduce the Linear Decay Label Smoothing Algorithm (LDLA), which progressively reduces uncertainty throughout the training process. Experiment on the HotpotQA dataset demonstrates the effectiveness of our methods in enhancing performance and generalizability in multi-hop reasoning, achieving new state-of-the-art results on the leaderboard.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Zhangyue Yin (27 papers)
  2. Yuxin Wang (132 papers)
  3. Xiannian Hu (2 papers)
  4. Yiguang Wu (1 paper)
  5. Hang Yan (86 papers)
  6. Xinyu Zhang (296 papers)
  7. Zhao Cao (36 papers)
  8. Xuanjing Huang (287 papers)
  9. Xipeng Qiu (257 papers)
Citations (9)