Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Probing for targeted syntactic knowledge through grammatical error detection (2210.16228v1)

Published 28 Oct 2022 in cs.CL

Abstract: Targeted studies testing knowledge of subject-verb agreement (SVA) indicate that pre-trained LLMs encode syntactic information. We assert that if models robustly encode subject-verb agreement, they should be able to identify when agreement is correct and when it is incorrect. To that end, we propose grammatical error detection as a diagnostic probe to evaluate token-level contextual representations for their knowledge of SVA. We evaluate contextual representations at each layer from five pre-trained English LLMs: BERT, XLNet, GPT-2, RoBERTa, and ELECTRA. We leverage public annotated training data from both English second language learners and Wikipedia edits, and report results on manually crafted stimuli for subject-verb agreement. We find that masked LLMs linearly encode information relevant to the detection of SVA errors, while the autoregressive models perform on par with our baseline. However, we also observe a divergence in performance when probes are trained on different training sets, and when they are evaluated on different syntactic constructions, suggesting the information pertaining to SVA error detection is not robustly encoded.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Christopher Davis (27 papers)
  2. Christopher Bryant (12 papers)
  3. Andrew Caines (13 papers)
  4. Marek Rei (52 papers)
  5. Paula Buttery (15 papers)
Citations (3)