Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Assessing Step-by-Step Reasoning against Lexical Negation: A Case Study on Syllogism (2310.14868v1)

Published 23 Oct 2023 in cs.CL

Abstract: LLMs take advantage of step-by-step reasoning instructions, e.g., chain-of-thought (CoT) prompting. Building on this, their ability to perform CoT-style reasoning robustly is of interest from a probing perspective. In this study, we inspect the step-by-step reasoning ability of LLMs with a focus on negation, which is a core linguistic phenomenon that is difficult to process. In particular, we introduce several controlled settings (e.g., reasoning in case of fictional entities) to evaluate the logical reasoning abilities of the models. We observed that dozens of modern LLMs were not robust against lexical negation (e.g., plausible ->implausible) when performing CoT-style reasoning, and the results highlight unique limitations in each LLM family.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Mengyu Ye (2 papers)
  2. Tatsuki Kuribayashi (31 papers)
  3. Jun Suzuki (86 papers)
  4. Goro Kobayashi (7 papers)
  5. Hiroaki Funayama (5 papers)
Citations (7)