Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Multilingual Transformer Encoders: a Word-Level Task-Agnostic Evaluation (2207.09076v1)

Published 19 Jul 2022 in cs.CL

Abstract: Some Transformer-based models can perform cross-lingual transfer learning: those models can be trained on a specific task in one language and give relatively good results on the same task in another language, despite having been pre-trained on monolingual tasks only. But, there is no consensus yet on whether those transformer-based models learn universal patterns across languages. We propose a word-level task-agnostic method to evaluate the alignment of contextualized representations built by such models. We show that our method provides more accurate translated word pairs than previous methods to evaluate word-level alignment. And our results show that some inner layers of multilingual Transformer-based models outperform other explicitly aligned representations, and even more so according to a stricter definition of multilingual alignment.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Félix Gaschi (5 papers)
  2. François Plesse (3 papers)
  3. Parisa Rastin (4 papers)
  4. Yannick Toussaint (5 papers)
Citations (6)

Summary

We haven't generated a summary for this paper yet.