Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

What do Transformers Know about Government? (2404.14270v1)

Published 22 Apr 2024 in cs.CL and cs.LG

Abstract: This paper investigates what insights about linguistic features and what knowledge about the structure of natural language can be obtained from the encodings in transformer LLMs.In particular, we explore how BERT encodes the government relation between constituents in a sentence. We use several probing classifiers, and data from two morphologically rich languages. Our experiments show that information about government is encoded across all transformer layers, but predominantly in the early layers of the model. We find that, for both languages, a small number of attention heads encode enough information about the government relations to enable us to train a classifier capable of discovering new, previously unknown types of government, never seen in the training data. Currently, data is lacking for the research community working on grammatical constructions, and government in particular. We release the Government Bank -- a dataset defining the government relations for thousands of lemmas in the languages in our experiments.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Jue Hou (35 papers)
  2. Anisia Katinskaia (7 papers)
  3. Lari Kotilainen (1 paper)
  4. Sathianpong Trangcasanchai (1 paper)
  5. Anh-Duc Vu (6 papers)
  6. Roman Yangarber (11 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.

HackerNews