Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Probing for Understanding of English Verb Classes and Alternations in Large Pre-trained Language Models (2209.04811v1)

Published 11 Sep 2022 in cs.CL

Abstract: We investigate the extent to which verb alternation classes, as described by Levin (1993), are encoded in the embeddings of Large Pre-trained LLMs (PLMs) such as BERT, RoBERTa, ELECTRA, and DeBERTa using selectively constructed diagnostic classifiers for word and sentence-level prediction tasks. We follow and expand upon the experiments of Kann et al. (2019), which aim to probe whether static embeddings encode frame-selectional properties of verbs. At both the word and sentence level, we find that contextual embeddings from PLMs not only outperform non-contextual embeddings, but achieve astonishingly high accuracies on tasks across most alternation classes. Additionally, we find evidence that the middle-to-upper layers of PLMs achieve better performance on average than the lower layers across all probing tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. David K. Yi (1 paper)
  2. James V. Bruno (1 paper)
  3. Jiayu Han (18 papers)
  4. Peter Zukerman (2 papers)
  5. Shane Steinert-Threlkeld (20 papers)
Citations (1)