2000 character limit reached
What makes multilingual BERT multilingual? (2010.10938v1)
Published 20 Oct 2020 in cs.CL and cs.LG
Abstract: Recently, multilingual BERT works remarkably well on cross-lingual transfer tasks, superior to static non-contextualized word embeddings. In this work, we provide an in-depth experimental study to supplement the existing literature of cross-lingual ability. We compare the cross-lingual ability of non-contextualized and contextualized representation model with the same data. We found that datasize and context window size are crucial factors to the transferability.
- Chi-Liang Liu (9 papers)
- Tsung-Yuan Hsu (6 papers)
- Yung-Sung Chuang (37 papers)
- Hung-yi Lee (325 papers)