Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Understanding the Capabilities and Limitations of Large Language Models for Cultural Commonsense (2405.04655v1)

Published 7 May 2024 in cs.CL

Abstract: LLMs have demonstrated substantial commonsense understanding through numerous benchmark evaluations. However, their understanding of cultural commonsense remains largely unexamined. In this paper, we conduct a comprehensive examination of the capabilities and limitations of several state-of-the-art LLMs in the context of cultural commonsense tasks. Using several general and cultural commonsense benchmarks, we find that (1) LLMs have a significant discrepancy in performance when tested on culture-specific commonsense knowledge for different cultures; (2) LLMs' general commonsense capability is affected by cultural context; and (3) The language used to query the LLMs can impact their performance on cultural-related tasks. Our study points to the inherent bias in the cultural understanding of LLMs and provides insights that can help develop culturally aware LLMs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Siqi Shen (29 papers)
  2. Lajanugen Logeswaran (30 papers)
  3. Moontae Lee (54 papers)
  4. Honglak Lee (174 papers)
  5. Soujanya Poria (138 papers)
  6. Rada Mihalcea (131 papers)
Citations (17)