Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
51 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Monolingual or Multilingual Instruction Tuning: Which Makes a Better Alpaca (2309.08958v2)

Published 16 Sep 2023 in cs.CL and cs.AI

Abstract: Foundational LLMs can be instruction-tuned to perform open-domain question answering, facilitating applications like chat assistants. While such efforts are often carried out in a single language, we empirically analyze cost-efficient strategies for multilingual scenarios. Our study employs the Alpaca dataset and machine translations of it to form multilingual data, which is then used to tune LLMs through either low-rank adaptation or full-parameter training. Under a controlled computation budget, comparisons show that multilingual tuning is on par or better than tuning a model for each language. Furthermore, multilingual tuning with downsampled data can be as powerful and more robust. Our findings serve as a guide for expanding language support through instruction tuning.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Pinzhen Chen (27 papers)
  2. Shaoxiong Ji (39 papers)
  3. Nikolay Bogoychev (17 papers)
  4. Barry Haddow (59 papers)
  5. Kenneth Heafield (24 papers)
  6. Andrey Kutuzov (41 papers)
Citations (35)