OMGEval: An Open Multilingual Generative Evaluation Benchmark for Large Language Models (2402.13524v1)
Abstract: Modern LLMs should generally benefit individuals from various cultural backgrounds around the world. However, most recent advanced generative evaluation benchmarks tailed for LLMs mainly focus on English. To this end, we introduce OMGEval, the first Open-source Multilingual Generative test set that can assess the capability of LLMs in different languages. For each language, OMGEval provides 804 open-ended questions, covering a wide range of important capabilities of LLMs, such as general knowledge, logical reasoning, and so on. Each question is rigorously verified by human annotators. Notably, to sufficiently reflect the compatibility of LLMs in different cultural backgrounds, we perform localization for each non-English language. Specifically, the current version of OMGEval includes 5 languages (i.e., Zh, Ru, Fr, Es, Ar). Following AlpacaEval, we employ GPT-4 as the adjudicator to automatically score different model outputs, which is shown closely related to human evaluation. We evaluate several representative multilingual LLMs on the proposed OMGEval, which we believe will provide a valuable reference for the community to further understand and improve the multilingual capability of LLMs. OMGEval is available at https://github.com/blcuicall/OMGEval.
- Yang Liu (2253 papers)
- Meng Xu (52 papers)
- Shuo Wang (382 papers)
- Liner Yang (22 papers)
- Haoyu Wang (309 papers)
- Zhenghao Liu (77 papers)
- Cunliang Kong (16 papers)
- Yun Chen (134 papers)
- Maosong Sun (337 papers)
- Erhong Yang (16 papers)