Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Causal Analysis of Syntactic Agreement Mechanisms in Neural Language Models (2106.06087v3)

Published 10 Jun 2021 in cs.CL

Abstract: Targeted syntactic evaluations have demonstrated the ability of LLMs to perform subject-verb agreement given difficult contexts. To elucidate the mechanisms by which the models accomplish this behavior, this study applies causal mediation analysis to pre-trained neural LLMs. We investigate the magnitude of models' preferences for grammatical inflections, as well as whether neurons process subject-verb agreement similarly across sentences with different syntactic structures. We uncover similarities and differences across architectures and model sizes -- notably, that larger models do not necessarily learn stronger preferences. We also observe two distinct mechanisms for producing subject-verb agreement depending on the syntactic structure of the input sentence. Finally, we find that LLMs rely on similar sets of neurons when given sentences with similar syntactic structure.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Matthew Finlayson (11 papers)
  2. Aaron Mueller (35 papers)
  3. Sebastian Gehrmann (48 papers)
  4. Stuart Shieber (6 papers)
  5. Tal Linzen (73 papers)
  6. Yonatan Belinkov (111 papers)
Citations (88)