Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Heads-up! Unsupervised Constituency Parsing via Self-Attention Heads (2010.09517v1)

Published 19 Oct 2020 in cs.CL

Abstract: Transformer-based pre-trained LLMs (PLMs) have dramatically improved the state of the art in NLP across many tasks. This has led to substantial interest in analyzing the syntactic knowledge PLMs learn. Previous approaches to this question have been limited, mostly using test suites or probes. Here, we propose a novel fully unsupervised parsing approach that extracts constituency trees from PLM attention heads. We rank transformer attention heads based on their inherent properties, and create an ensemble of high-ranking heads to produce the final tree. Our method is adaptable to low-resource languages, as it does not rely on development sets, which can be expensive to annotate. Our experiments show that the proposed method often outperform existing approaches if there is no development set present. Our unsupervised parser can also be used as a tool to analyze the grammars PLMs learn implicitly. For this, we use the parse trees induced by our method to train a neural PCFG and compare it to a grammar derived from a human-annotated treebank.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Bowen Li (166 papers)
  2. Taeuk Kim (38 papers)
  3. Reinald Kim Amplayo (28 papers)
  4. Frank Keller (45 papers)
Citations (16)

Summary

We haven't generated a summary for this paper yet.