Papers
Topics
Authors
Recent
2000 character limit reached

CMoralEval: A Moral Evaluation Benchmark for Chinese Large Language Models (2408.09819v1)

Published 19 Aug 2024 in cs.CL and cs.AI

Abstract: What a LLM would respond in ethically relevant context? In this paper, we curate a large benchmark CMoralEval for morality evaluation of Chinese LLMs. The data sources of CMoralEval are two-fold: 1) a Chinese TV program discussing Chinese moral norms with stories from the society and 2) a collection of Chinese moral anomies from various newspapers and academic papers on morality. With these sources, we aim to create a moral evaluation dataset characterized by diversity and authenticity. We develop a morality taxonomy and a set of fundamental moral principles that are not only rooted in traditional Chinese culture but also consistent with contemporary societal norms. To facilitate efficient construction and annotation of instances in CMoralEval, we establish a platform with AI-assisted instance generation to streamline the annotation process. These help us curate CMoralEval that encompasses both explicit moral scenarios (14,964 instances) and moral dilemma scenarios (15,424 instances), each with instances from different data sources. We conduct extensive experiments with CMoralEval to examine a variety of Chinese LLMs. Experiment results demonstrate that CMoralEval is a challenging benchmark for Chinese LLMs. The dataset is publicly available at \url{https://github.com/tjunlp-lab/CMoralEval}.

Citations (2)

Summary

  • The paper presents an innovative benchmark evaluating Chinese LLMs' moral reasoning through explicit moral scenarios and dilemmas derived from TV programs and academic sources.
  • It employs a comprehensive moral taxonomy covering familial, social, professional, Internet, and personal ethics, anchored in traditional Chinese principles.
  • Experimental results indicate many LLMs perform near random guessing, highlighting significant challenges in aligning AI with nuanced ethical standards.

CMoralEval: A Moral Evaluation Benchmark for Chinese LLMs

CMoralEval serves as an innovative benchmarking tool to assess the moral reasoning capabilities of Chinese LLMs. In response to the growing need for ethically aligned AI systems, CMoralEval offers a structured approach to evaluate Chinese LLMs against both explicit moral scenarios and moral dilemma scenarios. The development of this benchmark reflects a significant effort to integrate traditional Chinese moral principles with contemporary societal norms.

Methodology and Dataset Composition

CMoralEval is derived from two major sources: a Chinese TV program focusing on moral norms and a curated collection of Chinese moral anomies from newspapers and academic papers. The dataset consists of 30,388 instances, split into explicit moral scenarios (EMS) and moral dilemma scenarios (MDS), each composed of narratives designed to test various dimensions of morality.

The fundamental moral taxonomy incorporates familial morality, social morality, professional ethics, Internet ethics, and personal morality. This comprehensive classification ensures that ethical assessments span a wide range of cultural contexts and individual scenarios. A set of five fundamental moral principles, rooted in traditional Chinese values and Confucian moral theory, underpin these categories, providing clear criteria for evaluating moral alignment in the dataset. Figure 1

Figure 1: The benchmark workflow highlights the collection of questions from TV programs and moral anomies, essential for compiling CMoralEval.

Experimental Evaluation

Extensive experiments were conducted using CMoralEval on 26 different Chinese LLMs to evaluate their moral reasoning capabilities. Models were tested under zero-shot and few-shot settings to provide a detailed analysis of their ethical alignment capacities. The results indicate that CMoralEval presents a challenging benchmark, with many models performing near random guessing levels for moral discernment. Figure 2

Figure 2: Few-shot performance across subdivisions of CMoralEval, showcasing LLMs' handling of moral scenarios.

It was observed that larger models generally performed better in scenarios involving personal and familial morality, potentially due to more comprehensive training data that captures these nuances. Models tuned with reinforcement learning (RLHF) did not consistently outperform others, suggesting the complexity of tuning LLMs for moral tasks. Figure 3

Figure 3: Few-shot results across moral categories confirming Yi-34B-Chat's prominence in familial and personal morality contexts.

Implications and Future Prospects

The implications of CMoralEval extend beyond benchmarking; it is an essential step toward understanding how current LLMs align with ethical standards. The dataset exposes significant gaps in moral reasoning, necessitating advanced methodologies for enhancing LLM ethical training. Future developments may include incorporating moral context understanding and detailed ethical decision-making frameworks tailored for AI systems to address these deficiencies. Figure 4

Figure 4: Analysis of LLM performance on single-category versus multi-category questions in CMoralEval.

Conclusion

CMoralEval offers a novel and comprehensive approach to evaluating the moral reasoning capabilities of Chinese LLMs. While results show considerable room for improvement in LLMs' moral sensitivity, CMoralEval sets the path for further advancements in AI ethics within culturally nuanced frameworks. This benchmark underscores the need for continued exploration and refinement in AI moral reasoning and alignment to societal norms. Figure 5

Figure 5: Few-shot performance with controlling variables highlighting the inconsistencies in LLM ethical judgments.

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Github Logo Streamline Icon: https://streamlinehq.com