Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Cross-lingual Dependency Parsing with Unlabeled Auxiliary Languages (1909.09265v1)

Published 20 Sep 2019 in cs.CL

Abstract: Cross-lingual transfer learning has become an important weapon to battle the unavailability of annotated resources for low-resource languages. One of the fundamental techniques to transfer across languages is learning \emph{language-agnostic} representations, in the form of word embeddings or contextual encodings. In this work, we propose to leverage unannotated sentences from auxiliary languages to help learning language-agnostic representations. Specifically, we explore adversarial training for learning contextual encoders that produce invariant representations across languages to facilitate cross-lingual transfer. We conduct experiments on cross-lingual dependency parsing where we train a dependency parser on a source language and transfer it to a wide range of target languages. Experiments on 28 target languages demonstrate that adversarial training significantly improves the overall transfer performances under several different settings. We conduct a careful analysis to evaluate the language-agnostic representations resulted from adversarial training.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Wasi Uddin Ahmad (41 papers)
  2. Zhisong Zhang (31 papers)
  3. Xuezhe Ma (50 papers)
  4. Kai-Wei Chang (292 papers)
  5. Nanyun Peng (205 papers)
Citations (28)

Summary

We haven't generated a summary for this paper yet.