Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Køpsala: Transition-Based Graph Parsing via Efficient Training and Effective Encoding (2005.12094v2)

Published 25 May 2020 in cs.CL

Abstract: We present K{\o}psala, the Copenhagen-Uppsala system for the Enhanced Universal Dependencies Shared Task at IWPT 2020. Our system is a pipeline consisting of off-the-shelf models for everything but enhanced graph parsing, and for the latter, a transition-based graph parser adapted from Che et al. (2019). We train a single enhanced parser model per language, using gold sentence splitting and tokenization for training, and rely only on tokenized surface forms and multilingual BERT for encoding. While a bug introduced just before submission resulted in a severe drop in precision, its post-submission fix would bring us to 4th place in the official ranking, according to average ELAS. Our parser demonstrates that a unified pipeline is effective for both Meaning Representation Parsing and Enhanced Universal Dependencies.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Daniel Hershcovich (50 papers)
  2. Miryam de Lhoneux (29 papers)
  3. Artur Kulmizev (11 papers)
  4. Elham Pejhan (1 paper)
  5. Joakim Nivre (30 papers)
Citations (6)

Summary

We haven't generated a summary for this paper yet.