Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Zero-Shot Dependency Parsing with Worst-Case Aware Automated Curriculum Learning (2203.08555v1)

Published 16 Mar 2022 in cs.CL

Abstract: Large multilingual pretrained LLMs such as mBERT and XLM-RoBERTa have been found to be surprisingly effective for cross-lingual transfer of syntactic parsing models (Wu and Dredze 2019), but only between related languages. However, source and training languages are rarely related, when parsing truly low-resource languages. To close this gap, we adopt a method from multi-task learning, which relies on automated curriculum learning, to dynamically optimize for parsing performance on outlier languages. We show that this approach is significantly better than uniform and size-proportional sampling in the zero-shot setting.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Miryam de Lhoneux (29 papers)
  2. Sheng Zhang (212 papers)
  3. Anders Søgaard (121 papers)
Citations (10)

Summary

We haven't generated a summary for this paper yet.