Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

On Commonsense Cues in BERT for Solving Commonsense Tasks (2008.03945v3)

Published 10 Aug 2020 in cs.CL

Abstract: BERT has been used for solving commonsense tasks such as CommonsenseQA. While prior research has found that BERT does contain commonsense information to some extent, there has been work showing that pre-trained models can rely on spurious associations (e.g., data bias) rather than key cues in solving sentiment classification and other problems. We quantitatively investigate the presence of structural commonsense cues in BERT when solving commonsense tasks, and the importance of such cues for the model prediction. Using two different measures, we find that BERT does use relevant knowledge for solving the task, and the presence of commonsense knowledge is positively correlated to the model accuracy.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Leyang Cui (50 papers)
  2. Sijie Cheng (23 papers)
  3. Yu Wu (196 papers)
  4. Yue Zhang (620 papers)
Citations (13)

Summary

We haven't generated a summary for this paper yet.