Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
51 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

An Empirical Study on Cross-lingual Vocabulary Adaptation for Efficient Language Model Inference (2402.10712v3)

Published 16 Feb 2024 in cs.CL and cs.AI

Abstract: The development of state-of-the-art generative LLMs disproportionately relies on English-centric tokenizers, vocabulary and pre-training data. Despite the fact that some LLMs have multilingual capabilities, recent studies have shown that their inference efficiency deteriorates when generating text in languages other than English. This results in increased inference time and costs. Cross-lingual vocabulary adaptation (CVA) methods have been proposed for adapting models to a target language aiming to improve downstream performance. However, the effectiveness of these methods on increasing inference efficiency of generative LLMs has yet to be explored. In this paper, we perform an empirical study of five CVA methods on four generative LLMs (including monolingual and multilingual models) across four typologically-diverse languages and four natural language understanding tasks. We find that CVA substantially contributes to LLM inference speedups of up to 271.5\%. We also show that adapting LLMs that have been pre-trained on more balanced multilingual data results in downstream performance comparable to the original models.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Atsuki Yamaguchi (11 papers)
  2. Aline Villavicencio (31 papers)
  3. Nikolaos Aletras (72 papers)
Citations (3)
X Twitter Logo Streamline Icon: https://streamlinehq.com