NatLogAttack: A Framework for Attacking Natural Language Inference Models with Natural Logic (2307.02849v2)
Abstract: Reasoning has been a central topic in artificial intelligence from the beginning. The recent progress made on distributed representation and neural networks continues to improve the state-of-the-art performance of natural language inference. However, it remains an open question whether the models perform real reasoning to reach their conclusions or rely on spurious correlations. Adversarial attacks have proven to be an important tool to help evaluate the Achilles' heel of the victim models. In this study, we explore the fundamental problem of developing attack models based on logic formalism. We propose NatLogAttack to perform systematic attacks centring around natural logic, a classical logic formalism that is traceable back to Aristotle's syllogism and has been closely developed for natural language inference. The proposed framework renders both label-preserving and label-flipping attacks. We show that compared to the existing attack models, NatLogAttack generates better adversarial examples with fewer visits to the victim models. The victim models are found to be more vulnerable under the label-flipping setting. NatLogAttack provides a tool to probe the existing and future NLI models' capacity from a key viewpoint and we hope more logic-based attacks will be further explored for understanding the desired property of reasoning.
- Generating natural language adversarial examples. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 2890–2896, Brussels, Belgium. Association for Computational Linguistics.
- Gabor Angeli and Christopher D Manning. 2014. Naturalli: Natural logic inference for common sense reasoning. In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), Doha, Qatar.
- Combining natural logic and shallow reasoning for question answering. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (ACL), Berlin, Germany.
- Don’t take the premise for granted: Mitigating artifacts in natural language inference. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (ACL), pages 877–891.
- A large annotated corpus for learning natural language inference. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 632–642, Lisbon, Portugal. Association for Computational Linguistics.
- Neural natural language inference models enhanced with external knowledge. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (ACL), Melbourne, Australia.
- Enhanced lstm for natural language inference. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (ACL), Vancouver. ACL.
- Zeming Chen and Qiyue Gao. 2021. Monotonicity marking from universal dependency trees. In Proceedings of the 14th International Conference on Computational Semantics (IWCS), pages 121–131.
- Neurallog: Natural language inference with joint neural and logical reasoning. In Proceedings of* SEM 2021: The Tenth Joint Conference on Lexical and Computational Semantics, pages 78–88.
- The PASCAL recognising textual entailment challenge. In Proceedings of the First international conference on Machine Learning Challenges: evaluating Predictive Uncertainty Visual Object Classification, and Recognizing Textual Entailment.
- Neuro-symbolic= neural+ logical+ probabilistic. In NeSy’19@ IJCAI, the 14th International Workshop on Neural-Symbolic Learning and Reasoning, Macao, China.
- BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota. Association for Computational Linguistics.
- HowNet and its computation of meaning. In Coling 2010: Demonstrations, pages 53–56, Beijing, China. Coling 2010 Organizing Committee.
- HotFlip: White-box adversarial examples for text classification. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 31–36, Melbourne, Australia. Association for Computational Linguistics.
- Richard Evans and Edward Grefenstette. 2018. Learning explanatory rules from noisy data. In Journal of Artificial Intelligence Research (JAIR), volume 61, pages 1–64.
- Neuro-symbolic natural logic with introspective revision for natural language inference. Transactions of the Association for Computational Linguistics (TACL), 10:240–256.
- Exploring end-to-end differentiable natural logic modeling. In Proceedings of the 28th International Conference on Computational Linguistics (COLING), pages 1172–1185.
- Neural-symbolic learning and reasoning: contributions and challenges. In 2015 AAAI Spring Symposium Series.
- Neural natural language inference models partially embed theories of lexical entailment and negation. In Proceedings of the Third BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP, pages 163–173.
- Breaking NLI systems with sentences that require simple lexical inferences. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 650–655, Melbourne, Australia. Association for Computational Linguistics.
- Explaining and harnessing adversarial examples. arXiv preprint arXiv:1412.6572.
- Monalog: a lightweight system for natural language inference based on monotonicity.
- Hai Hu and Larry Moss. 2018. Polarity computations in flexible categorial grammar. In Proceedings of the Seventh Joint Conference on Lexical and Computational Semantics, pages 124–129.
- Thomas F Icard. 2012. Inclusion and exclusion in natural language. Studia Logica.
- Thomas F Icard and Lawrence S Moss. 2014. Recent progress on monotonicity. In Linguistic Issues in Language Technology. Citeseer.
- Adrian Iftene and Alexandra Balahur-Dobrescu. 2007. Proceedings of the ACL-PASCAL Workshop on Textual Entailment and Paraphrasing. Prague, Czech.
- Adversarial example generation with syntactically controlled paraphrase networks. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 1875–1885.
- Robin Jia and Percy Liang. 2017. Adversarial examples for evaluating reading comprehension systems. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 2021–2031, Copenhagen, Denmark. Association for Computational Linguistics.
- Is bert really robust? a strong baseline for natural language attack on text classification and entailment. In Proceedings of the AAAI conference on artificial intelligence, volume 34, pages 8018–8025.
- Adventure: Adversarial training for textual entailment with knowledge-guided examples. In The 56th Annual Meeting of the Association for Computational Linguistics (ACL), Melbourne, Australia.
- Adversarial examples in the physical world. In Artificial intelligence safety and security, pages 99–112. Chapman and Hall/CRC.
- George Lakoff. 1970. Linguistics and natural logic. Synthese, 22(1-2):151–271.
- Contextualized perturbation for textual adversarial attack. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics (NAACL).
- BERT-ATTACK: Adversarial attack against BERT using BERT. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 6193–6202, Online. Association for Computational Linguistics.
- Deep text classification can be fooled. In Proceedings of the 27th International Joint Conference on Artificial Intelligence (IJCAI), pages 4208–4215.
- Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692.
- Bill MacCartney. 2009. Natural Language Inference. Ph.D. thesis, Stanford University.
- Bill MacCartney and Christopher D Manning. 2009. An extended model of natural logic. In Proceedings of the 8th international conference on computational semantics (IWCS), Stroudsburg, United States.
- The Neuro-Symbolic Concept Learner: Interpreting Scenes, Words, and Sentences From Natural Supervision. In Proceedings of the 7th International Conference on Learning Representations (ICLR), New Orleans, USA.
- A sick cure for the evaluation of compositional distributional semantic models. In Proceedings of the 9th International Conference on Language Resources and Evaluation (LREC), Reykjavik, Iceland.
- Right for the wrong reasons: Diagnosing syntactic heuristics in natural language inference. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (ACL), pages 3428–3448.
- George A Miller. 1995. Wordnet: a lexical database for english. Communications of the ACM, 38(11):39–41.
- Pasquale Minervini and Sebastian Riedel. 2018. Adversarially regularising neural nli models to integrate logical background knowledge. In Proceedings of the 22nd Conference on Computational Natural Language Learning (CoNLL), pages 65–74.
- Textattack: A framework for adversarial attacks, data augmentation, and adversarial training in nlp. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 119–126.
- Computing relative polarity for textual inference. In Proceedings of the 5th international workshop on inference in computational semantics, Buxton, England.
- Barbara Partee. 1995. Lexical semantics and compositionality. Invitation to Cognitive Science.
- Conditionally adaptive multi-task learning: Improving transfer learning in nlp using fewer parameters & less data. In International Conference on Learning Representations (ICLR).
- Collecting diverse natural language inference problems for sentence representation evaluation. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 67–81, Brussels, Belgium. Association for Computational Linguistics.
- Language models are unsupervised multitask learners. OpenAI Blog, 1(8):9.
- Generating natural language adversarial examples through probability weighted word saliency. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (ACL), pages 1085–1097.
- Probing natural language inference models through semantic fragments. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI), New York, USA.
- Tim Rocktäschel and Sebastian Riedel. 2017. End-to-end differentiable proving. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NeurIPS), Long Beach, USA.
- Masked language model scoring. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 2699–2712, Online. Association for Computational Linguistics.
- Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter. the 5th Workshop on Energy Efficient Machine Learning and Cognitive Computing @NeurIPS.
- Interpretable adversarial perturbation in input embedding space for text. In Proceedings of the 27th International Joint Conference on Artificial Intelligence (IJCAI), pages 4323–4330.
- Fundamental tradeoffs between invariance and sensitivity to adversarial perturbations. In International Conference on Machine Learning, pages 9561–9571. PMLR.
- Víctor Manuel Sánchez Valencia. 1991. Studies on natural logic and categorial grammar. Universiteit van Amsterdam.
- Johan van Benthem. 1988. The semantics of variety in categorial grammar. Categorial grammar.
- Johan Van Benthem. 1995. Language in Action: categories, lambdas and dynamic logic. MIT Press.
- Johan Van Benthem et al. 1986. Essays in logical semantics. Springer.
- Universal adversarial triggers for attacking and analyzing NLP. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 2153–2162, Hong Kong, China. Association for Computational Linguistics.
- Nlprolog: Reasoning with weak unification for question answering in natural language. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (ACL), Austin, Texas, United States.
- A broad-coverage challenge corpus for sentence understanding through inference. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 1112–1122, New Orleans, Louisiana. Association for Computational Linguistics.
- Do neural models learn systematicity of monotonicity inference in natural language? In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics (ACL), pages 6105–6117.
- Can neural networks understand monotonicity reasoning? In Proceedings of the 2019 ACL Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, Austin, Texas, United States.
- Help: A dataset for identifying shortcomings of neural models in monotonicity reasoning. In Proceedings of the Eighth Joint Conference on Lexical and Computational Semantics (*SEM), Minneapolis, Minnesota, USA.
- Differentiable learning of logical rules for knowledge base reasoning. In Advances in Neural Information Processing Systems, pages 2319–2328.
- Word-level textual adversarial attacking as combinatorial optimization. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics (ACL), pages 6066–6080.
- Openattack: An open-source textual adversarial attack toolkit. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing: System Demonstrations, pages 363–371.
- Generating fluent adversarial examples for natural languages. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (ACL), pages 5564–5569, Florence, Italy. Association for Computational Linguistics.
- Semantics-aware bert for language understanding. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 9628–9635.
- Generating natural adversarial examples. In Proceedings of the 6th International Conference on Learning Representations (ICLR).
- Zi'ou Zheng (4 papers)
- Xiaodan Zhu (94 papers)