Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

InstructAlign: High-and-Low Resource Language Alignment via Continual Crosslingual Instruction Tuning (2305.13627v2)

Published 23 May 2023 in cs.CL and cs.AI

Abstract: LLMs that are tuned with instructions have demonstrated remarkable capabilities in various tasks and languages. However, their ability to generalize to underrepresented languages is limited due to the scarcity of available data. Additionally, directly adapting new languages to instruction-tuned LLMs can result in catastrophic forgetting, which leads to the loss of multitasking ability. To address this issue, we propose InstructAlign which uses continual crosslingual instruction tuning to enable LLMs to align new unseen languages with previously learned high-resource languages. Our results demonstrate the effectiveness of InstructAlign in enabling the model to understand low-resource languages with limited parallel data while preventing catastrophic forgetting. Our work contributes to the advancement of language adaptation methods, particularly for adapting instruction-tuned LLMs to underrepresented languages. Our code is released on https://github.com/HLTCHKUST/InstructAlign

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Samuel Cahyawijaya (75 papers)
  2. Holy Lovenia (30 papers)
  3. Tiezheng Yu (29 papers)
  4. Willy Chung (10 papers)
  5. Pascale Fung (150 papers)
Citations (11)
Github Logo Streamline Icon: https://streamlinehq.com