Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Comparison of Architectures and Pretraining Methods for Contextualized Multilingual Word Embeddings (1912.10169v1)

Published 15 Dec 2019 in cs.CL

Abstract: The lack of annotated data in many languages is a well-known challenge within the field of multilingual NLP. Therefore, many recent studies focus on zero-shot transfer learning and joint training across languages to overcome data scarcity for low-resource languages. In this work we (i) perform a comprehensive comparison of state-ofthe-art multilingual word and sentence encoders on the tasks of named entity recognition (NER) and part of speech (POS) tagging; and (ii) propose a new method for creating multilingual contextualized word embeddings, compare it to multiple baselines and show that it performs at or above state-of-theart level in zero-shot transfer settings. Finally, we show that our method allows for better knowledge sharing across languages in a joint training setting.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Niels van der Heijden (4 papers)
  2. Samira Abnar (19 papers)
  3. Ekaterina Shutova (52 papers)
Citations (16)