Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Structural Persistence in Language Models: Priming as a Window into Abstract Language Representations (2109.14989v2)

Published 30 Sep 2021 in cs.CL

Abstract: We investigate the extent to which modern, neural LLMs are susceptible to structural priming, the phenomenon whereby the structure of a sentence makes the same structure more probable in a follow-up sentence. We explore how priming can be used to study the potential of these models to learn abstract structural information, which is a prerequisite for good performance on tasks that require natural language understanding skills. We introduce a novel metric and release Prime-LM, a large corpus where we control for various linguistic factors which interact with priming strength. We find that Transformer models indeed show evidence of structural priming, but also that the generalisations they learned are to some extent modulated by semantic information. Our experiments also show that the representations acquired by the models may not only encode abstract sequential structure but involve certain level of hierarchical syntactic information. More generally, our study shows that the priming paradigm is a useful, additional tool for gaining insights into the capacities of LLMs and opens the door to future priming-based investigations that probe the model's internal states.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Arabella Sinclair (6 papers)
  2. Jaap Jumelet (25 papers)
  3. Willem Zuidema (32 papers)
  4. Raquel Fernández (52 papers)
Citations (35)

Summary

We haven't generated a summary for this paper yet.