Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Linguistic Interpretability of Transformer-based Language Models: a systematic review (2504.08001v1)

Published 9 Apr 2025 in cs.CL

Abstract: LLMs based on the Transformer architecture achieve excellent results in many language-related tasks, such as text classification or sentiment analysis. However, despite the architecture of these models being well-defined, little is known about how their internal computations help them achieve their results. This renders these models, as of today, a type of 'black box' systems. There is, however, a line of research -- 'interpretability' -- aiming to learn how information is encoded inside these models. More specifically, there is work dedicated to studying whether Transformer-based models possess knowledge of linguistic phenomena similar to human speakers -- an area we call 'linguistic interpretability' of these models. In this survey we present a comprehensive analysis of 160 research works, spread across multiple languages and models -- including multilingual ones -- that attempt to discover linguistic information from the perspective of several traditional Linguistics disciplines: Syntax, Morphology, Lexico-Semantics and Discourse. Our survey fills a gap in the existing interpretability literature, which either not focus on linguistic knowledge in these models or present some limitations -- e.g. only studying English-based models. Our survey also focuses on Pre-trained LLMs not further specialized for a downstream task, with an emphasis on works that use interpretability techniques that explore models' internal representations.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Miguel López-Otal (1 paper)
  2. Jorge Gracia (3 papers)
  3. Jordi Bernad (1 paper)
  4. Carlos Bobed (4 papers)
  5. Lucía Pitarch-Ballesteros (1 paper)
  6. Emma Anglés-Herrero (1 paper)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets