Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Methodology of Adapting Large English Language Models for Specific Cultural Contexts (2406.18192v2)

Published 26 Jun 2024 in cs.CL and cs.AI

Abstract: The rapid growth of LLMs(LLMs) has emerged as a prominent trend in the field of artificial intelligence. However, current state-of-the-art LLMs are predominantly based on English. They encounter limitations when directly applied to tasks in specific cultural domains, due to deficiencies in domain-specific knowledge and misunderstandings caused by differences in cultural values. To address this challenge, our paper proposes a rapid adaptation method for large models in specific cultural contexts, which leverages instruction-tuning based on specific cultural knowledge and safety values data. Taking Chinese as the specific cultural context and utilizing the LLaMA3-8B as the experimental English LLM, the evaluation results demonstrate that the adapted LLM significantly enhances its capabilities in domain-specific knowledge and adaptability to safety values, while maintaining its original expertise advantages.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Wenjing Zhang (28 papers)
  2. Siqi Xiao (2 papers)
  3. Xuejiao Lei (6 papers)
  4. Ning Wang (300 papers)
  5. Huazheng Zhang (1 paper)
  6. Meijuan An (4 papers)
  7. Bikun Yang (3 papers)
  8. Zhaoxiang Liu (54 papers)
  9. Kai Wang (624 papers)
  10. Shiguo Lian (54 papers)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets