Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
133 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Polyglot Contextual Representations Improve Crosslingual Transfer (1902.09697v2)

Published 26 Feb 2019 in cs.CL

Abstract: We introduce Rosita, a method to produce multilingual contextual word representations by training a single LLM on text from multiple languages. Our method combines the advantages of contextual word representations with those of multilingual representation learning. We produce LLMs from dissimilar language pairs (English/Arabic and English/Chinese) and use them in dependency parsing, semantic role labeling, and named entity recognition, with comparisons to monolingual and non-contextual variants. Our results provide further evidence for the benefits of polyglot learning, in which representations are shared across multiple languages.

Citations (68)

Summary

We haven't generated a summary for this paper yet.