Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
72 tokens/sec
GPT-4o
61 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

1+1>2: Can Large Language Models Serve as Cross-Lingual Knowledge Aggregators? (2406.14721v1)

Published 20 Jun 2024 in cs.CL

Abstract: LLMs have garnered significant attention due to their remarkable ability to process information across various languages. Despite their capabilities, they exhibit inconsistencies in handling identical queries in different languages, presenting challenges for further advancement. This paper introduces a method to enhance the multilingual performance of LLMs by aggregating knowledge from diverse languages. This approach incorporates a low-resource knowledge detector specific to a language, a language selection process, and mechanisms for answer replacement and integration. Our experiments demonstrate notable performance improvements, particularly in reducing language performance disparity. An ablation study confirms that each component of our method significantly contributes to these enhancements. This research highlights the inherent potential of LLMs to harmonize multilingual capabilities and offers valuable insights for further exploration.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Yue Huang (171 papers)
  2. Chenrui Fan (9 papers)
  3. Yuan Li (392 papers)
  4. Siyuan Wu (18 papers)
  5. Tianyi Zhou (172 papers)
  6. Xiangliang Zhang (131 papers)
  7. Lichao Sun (186 papers)
X Twitter Logo Streamline Icon: https://streamlinehq.com