2000 character limit reached
longhorns at DADC 2022: How many linguists does it take to fool a Question Answering model? A systematic approach to adversarial attacks (2206.14729v1)
Published 29 Jun 2022 in cs.CL, cs.AI, and cs.HC
Abstract: Developing methods to adversarially challenge NLP systems is a promising avenue for improving both model performance and interpretability. Here, we describe the approach of the team "longhorns" on Task 1 of the The First Workshop on Dynamic Adversarial Data Collection (DADC), which asked teams to manually fool a model on an Extractive Question Answering task. Our team finished first, with a model error rate of 62%. We advocate for a systematic, linguistically informed approach to formulating adversarial questions, and we describe the results of our pilot experiments, as well as our official submission.
- Venelin Kovatchev (12 papers)
- Trina Chatterjee (2 papers)
- Venkata S Govindarajan (6 papers)
- Jifan Chen (12 papers)
- Eunsol Choi (76 papers)
- Gabriella Chronis (2 papers)
- Anubrata Das (12 papers)
- Katrin Erk (23 papers)
- Matthew Lease (57 papers)
- Junyi Jessy Li (79 papers)
- Yating Wu (9 papers)
- Kyle Mahowald (40 papers)