2000 character limit reached
Specializing Multilingual Language Models: An Empirical Study (2106.09063v4)
Published 16 Jun 2021 in cs.CL
Abstract: Pretrained multilingual LLMs have become a common tool in transferring NLP capabilities to low-resource languages, often with adaptations. In this work, we study the performance, extensibility, and interaction of two such adaptations: vocabulary augmentation and script transliteration. Our evaluations on part-of-speech tagging, universal dependency parsing, and named entity recognition in nine diverse low-resource languages uphold the viability of these approaches while raising new questions around how to optimally adapt multilingual models to low-resource settings.
- Ethan C. Chau (5 papers)
- Noah A. Smith (224 papers)