Generating Diverse Negations from Affirmative Sentences (2411.00056v1)
Abstract: Despite the impressive performance of LLMs across various tasks, they often struggle with reasoning under negated statements. Negations are important in real-world applications as they encode negative polarity in verb phrases, clauses, or other expressions. Nevertheless, they are underrepresented in current benchmarks, which mainly include basic negation forms and overlook more complex ones, resulting in insufficient data for training a LLM. In this work, we propose NegVerse, a method that tackles the lack of negation datasets by producing a diverse range of negation types from affirmative sentences, including verbal, non-verbal, and affixal forms commonly found in English text. We provide new rules for masking parts of sentences where negations are most likely to occur, based on syntactic structure and use a frozen baseline LLM and prompt tuning to generate negated sentences. We also propose a filtering mechanism to identify negation cues and remove degenerate examples, producing a diverse range of meaningful perturbations. Our results show that NegVerse outperforms existing methods and generates negations with higher lexical similarity to the original sentences, better syntactic preservation and negation diversity. The code is available in https://github.com/DarianRodriguez/NegVerse
- Strong hallucinations from negation and how to fix them, 2024.
- Text summarization using large language models: A comparative study of mpt-7b-instruct, falcon-7b-instruct, and openai chat-gpt models, 2023.
- A large annotated corpus for learning natural language inference. In Lluís Màrquez, Chris Callison-Burch, and Jian Su, editors, Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pages 632–642, Lisbon, Portugal, September 2015. Association for Computational Linguistics.
- Language models are few-shot learners, 2020.
- SemEval-2017 task 1: Semantic textual similarity multilingual and crosslingual focused evaluation. In Steven Bethard, Marine Carpuat, Marianna Apidianaki, Saif M. Mohammad, Daniel Cer, and David Jurgens, editors, Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017), pages 1–14, Vancouver, Canada, August 2017. Association for Computational Linguistics.
- Language models can exploit cross-task in-context learning for data-scarce novel tasks, 2024.
- BERT: Pre-training of deep bidirectional transformers for language understanding. In Jill Burstein, Christy Doran, and Thamar Solorio, editors, Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota, June 2019. Association for Computational Linguistics.
- This is not a dataset: A large negation benchmark to challenge large language models, 2023.
- Exploring Human-Like Translation Strategy with Large Language Models. Transactions of the Association for Computational Linguistics, 12:229–246, 03 2024.
- TINA: Textual inference with negation augmentation. In Yoav Goldberg, Zornitsa Kozareva, and Yue Zhang, editors, Findings of the Association for Computational Linguistics: EMNLP 2022, pages 4086–4099, Abu Dhabi, United Arab Emirates, December 2022. Association for Computational Linguistics.
- Leveraging affirmative interpretations from negation improves natural language understanding, 2022.
- An analysis of negation in natural language understanding corpora, 2022.
- An analysis of natural language inference benchmarks through the lens of negation. In Bonnie Webber, Trevor Cohn, Yulan He, and Yang Liu, editors, Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 9106–9118, Online, November 2020. Association for Computational Linguistics.
- Understanding by understanding not: Modeling negation in language models, 2021.
- Can large language models truly understand prompts? a case study with negated prompts, 2022.
- BECEL: Benchmark for consistency evaluation of language models. In Nicoletta Calzolari, Chu-Ren Huang, Hansaem Kim, James Pustejovsky, Leo Wanner, Key-Sun Choi, Pum-Mo Ryu, Hsin-Hsi Chen, Lucia Donatelli, Heng Ji, Sadao Kurohashi, Patrizia Paggio, Nianwen Xue, Seokhwan Kim, Younggyun Hahm, Zhong He, Tony Kyungil Lee, Enrico Santus, Francis Bond, and Seung-Hoon Na, editors, Proceedings of the 29th International Conference on Computational Linguistics, pages 3680–3696, Gyeongju, Republic of Korea, October 2022. International Committee on Computational Linguistics.
- Evaluating open-domain question answering in the era of large language models, 2023.
- Negated and misprimed probes for pretrained language models: Birds can talk, but cannot fly, 2020.
- Negbert: A transfer learning approach for negation detection and scope resolution, 2020.
- Logical negation augmenting and debiasing for prompt-based methods, 2024.
- Roberta: A robustly optimized bert pretraining approach, 2019.
- Generate your counterfactuals: Towards controlled counterfactual generation for text, 2021.
- Recent advances in processing negation. Natural Language Engineering, 27(2):121–130, 2021.
- Mteb: Massive text embedding benchmark, 2023.
- Llms for generating and evaluating counterfactuals: A comprehensive study, 2024.
- R OpenAI et al. Gpt-4 technical report, 2024.
- A review of current trends, techniques, and challenges in large language models (llms). Applied Sciences, 2024.
- SemEval-2014 task 4: Aspect based sentiment analysis. In Preslav Nakov and Torsten Zesch, editors, Proceedings of the 8th International Workshop on Semantic Evaluation (SemEval 2014), pages 27–35, Dublin, Ireland, August 2014. Association for Computational Linguistics.
- Choice of plausible alternatives: An evaluation of commonsense causal reasoning. In AAAI Spring Symposium on Logical Formalizations of Commonsense Reasoning, 01 2011.
- Explaining nlp models via minimal contrastive editing (mice), 2021.
- CATfOOD: Counterfactual augmented training for improving out-of-domain performance and calibration. In Yvette Graham and Matthew Purver, editors, Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1876–1898, St. Julian’s, Malta, March 2024. Association for Computational Linguistics.
- A large language model screening tool to target patients for best practice alerts: Development and validation. JMIR Med Inform, 11:e49886, Nov 2023.
- Dependency parse reranking with rich subtree features. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 22:1208–1218, 2014.
- NLMs: Augmenting negation in language models. In Houda Bouamor, Juan Pino, and Kalika Bali, editors, Findings of the Association for Computational Linguistics: EMNLP 2023, pages 13104–13116, Singapore, December 2023. Association for Computational Linguistics.
- Recursive deep models for semantic compositionality over a sentiment treebank. In David Yarowsky, Timothy Baldwin, Anna Korhonen, Karen Livescu, and Steven Bethard, editors, Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, pages 1631–1642, Seattle, Washington, USA, October 2013. Association for Computational Linguistics.
- Christopher Toukmaji. Few-shot cross-lingual transfer for prompting large language models in low-resource languages, 2024.
- Llama 2: Open foundation and fine-tuned chat models, 2023.
- Crest: A joint framework for rationalization and counterfactual text generation, 2023.
- Language models are not naysayers: An analysis of language models on negation benchmarks, 2023.
- Not another negation benchmark: The NaN-NLI test suite for sub-clausal negation. In Yulan He, Heng Ji, Sujian Li, Yang Liu, and Chua-Hui Chang, editors, Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 883–894, Online only, November 2022. Association for Computational Linguistics.
- Building a dictionary of affixal negations. In Eduardo Blanco, Roser Morante, and Roser Saurí, editors, Proceedings of the Workshop on Extra-Propositional Aspects of Meaning in Computational Linguistics (ExProM), pages 49–56, Osaka, Japan, December 2016. The COLING 2016 Organizing Committee.
- Automated evaluation of personalized text generation using large language models, 2023.
- Polyjuice: Generating counterfactuals for explaining, evaluating, and improving models, 2021.
- SOME: Reference-less sub-metrics optimized for manual evaluations of grammatical error correction. In Donia Scott, Nuria Bel, and Chengqing Zong, editors, Proceedings of the 28th International Conference on Computational Linguistics, pages 6516–6522, Barcelona, Spain (Online), December 2020. International Committee on Computational Linguistics.
- Simple fast algorithms for the editing distance between trees and related problems. SIAM J. Comput., 18:1245–1262, 1989.
- Texygen: A benchmarking platform for text generation models, 2018.
- Darian Rodriguez Vasquez (1 paper)
- Afroditi Papadaki (5 papers)