Methodology of Adapting Large English Language Models for Specific Cultural Contexts (2406.18192v2)
Abstract: The rapid growth of LLMs(LLMs) has emerged as a prominent trend in the field of artificial intelligence. However, current state-of-the-art LLMs are predominantly based on English. They encounter limitations when directly applied to tasks in specific cultural domains, due to deficiencies in domain-specific knowledge and misunderstandings caused by differences in cultural values. To address this challenge, our paper proposes a rapid adaptation method for large models in specific cultural contexts, which leverages instruction-tuning based on specific cultural knowledge and safety values data. Taking Chinese as the specific cultural context and utilizing the LLaMA3-8B as the experimental English LLM, the evaluation results demonstrate that the adapted LLM significantly enhances its capabilities in domain-specific knowledge and adaptability to safety values, while maintaining its original expertise advantages.
- Wenjing Zhang (28 papers)
- Siqi Xiao (2 papers)
- Xuejiao Lei (6 papers)
- Ning Wang (300 papers)
- Huazheng Zhang (1 paper)
- Meijuan An (4 papers)
- Bikun Yang (3 papers)
- Zhaoxiang Liu (54 papers)
- Kai Wang (624 papers)
- Shiguo Lian (54 papers)