2000 character limit reached
When to Fold'em: How to answer Unanswerable questions (2105.00328v1)
Published 1 May 2021 in cs.CL
Abstract: We present 3 different question-answering models trained on the SQuAD2.0 dataset -- BIDAF, DocumentQA and ALBERT Retro-Reader -- demonstrating the improvement of LLMs in the past three years. Through our research in fine-tuning pre-trained models for question-answering, we developed a novel approach capable of achieving a 2% point improvement in SQuAD2.0 F1 in reduced training time. Our method of re-initializing select layers of a parameter-shared LLM is simple yet empirically powerful.
- Marshall Ho (2 papers)
- Zhipeng Zhou (32 papers)
- Judith He (1 paper)