Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
157 tokens/sec
GPT-4o
43 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Logical Negation Augmenting and Debiasing for Prompt-based Methods (2405.04872v1)

Published 8 May 2024 in cs.CL, cs.AI, and cs.LO

Abstract: Prompt-based methods have gained increasing attention on NLP and shown validity on many downstream tasks. Many works have focused on mining these methods' potential for knowledge extraction, but few explore their ability to make logical reasoning. In this work, we focus on the effectiveness of the prompt-based methods on first-order logical reasoning and find that the bottleneck lies in logical negation. Based on our analysis, logical negation tends to result in spurious correlations to negative answers, while propositions without logical negation correlate to positive answers. To solve the problem, we propose a simple but effective method, Negation Augmenting and Negation Debiasing (NAND), which introduces negative propositions to prompt-based methods without updating parameters. Specifically, these negative propositions can counteract spurious correlations by providing "not" for all instances so that models cannot make decisions only by whether expressions contain a logical negation. Experiments on three datasets show that NAND not only solves the problem of calibrating logical negation but also significantly enhances prompt-based methods of logical reasoning without model retraining.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (39)
  1. Language models are few-shot learners. In NeurIPS.
  2. Implicit premise generation with discourse-aware commonsense knowledge models. In EMNLP.
  3. Transformers as soft reasoners over language. In IJCAI.
  4. Template-based named entity recognition using BART. In ACL/IJCNLP.
  5. Ernest Davis. 2017. Logical formalizations of commonsense reasoning: A survey. J. Artif. Intell. Res.
  6. BERT: pre-training of deep bidirectional transformers for language understanding. In NAACL-HLT.
  7. Allyson Ettinger. 2020. What BERT is not: Lessons from a new suite of psycholinguistic diagnostics for language models. Trans. Assoc. Comput. Linguistics.
  8. Making pre-trained language models better few-shot learners. In ACL/IJCNLP.
  9. M. Gelfond and V. Lifschitz. 1991. Classical negation in logic pro-grams and deductive databases. New Generation Computing.
  10. Stack: Sentence ordering with temporal commonsense knowledge. In EMNLP.
  11. Annotation artifacts in natural language inference data. In NAACL-HLT.
  12. PTR: prompt tuning with rules for text classification. CoRR.
  13. Laurence R. Horn and Heinrich Wansing. 2020. Negation. In The Stanford Encyclopedia of Philosophy.
  14. Predicting the focus of negation: Model and error analysis. In ACL.
  15. Understanding by understanding not: Modeling negation in language models. In NAACL-HLT.
  16. What does BERT learn about the structure of language? In ACL.
  17. How can we know what language models know. Trans. Assoc. Comput. Linguistics.
  18. Nora Kassner and Hinrich Schütze. 2020. Negated and misprimed probes for pretrained language models: Birds can talk, but cannot fly. In ACL.
  19. Are prompt-based models clueless? In ACL.
  20. Aditya Khandelwal and Suraj Sawant. 2020. Negbert: A transfer learning approach for negation detection and scope resolution. In LREC.
  21. Generated knowledge prompting for commonsense reasoning. In ACL.
  22. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. CoRR.
  23. Roberta: A robustly optimized BERT pretraining approach. CoRR.
  24. Fantastically ordered prompts and where to find them: Overcoming few-shot prompt order sensitivity. In ACL.
  25. Language models as knowledge bases? In EMNLP-IJCNLP.
  26. Raymond Reiter. 1981. On closed world data bases. In Readings in Artificial Intelligence.
  27. Laria Reynolds and Kyle McDonell. 2021. Prompt programming for large language models: Beyond the few-shot paradigm. In CHI.
  28. Timo Schick and Hinrich Schütze. 2020. Rare words: A major problem for contextualized embeddings and how to fix it by attentive mimicking. In AAAI.
  29. Autoprompt: Eliciting knowledge from language models with automatically generated prompts. In EMNLP.
  30. Proofwriter: Generating implications, proofs, and abductive statements over natural language. In ACL/IJCNLP.
  31. olmpics - on what language model pre-training captures. Trans. Assoc. Comput. Linguistics.
  32. Diagnosing the first-order logical reasoning ability through logicnli. In EMNLP.
  33. Learning disentangled representations of negation and uncertainty. In ACL.
  34. Universal adversarial triggers for attacking and analyzing NLP. In EMNLP-IJCNLP.
  35. Can generative pre-trained language models serve as knowledge bases for closed-book qa? In ACL/IJCNLP.
  36. How does BERT capture semantics? A closer look at polysemous words. In BlackboxNLP@EMNLP.
  37. Reclor: A reading comprehension dataset requiring logical reasoning. In ICLR.
  38. Differentiable prompt makes pre-trained language models better few-shot learners. CoRR.
  39. Calibrate before use: Improving few-shot performance of language models. In ICML.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets