Transformers in the Service of Description Logic-based Contexts (2311.08941v3)
Abstract: Recent advancements in transformer-based models have initiated research interests in investigating their ability to learn to perform reasoning tasks. However, most of the contexts used for this purpose are in practice very simple: generated from short (fragments of) first-order logic sentences with only a few logical operators and quantifiers. In this work, we construct the natural language dataset, DELTA$_D$, using the description logic language $\mathcal{ALCQ}$. DELTA$_D$ contains 384K examples, and increases in two dimensions: i) reasoning depth, and ii) linguistic complexity. In this way, we systematically investigate the reasoning ability of a supervised fine-tuned DeBERTa-based model and of two LLMs (GPT-3.5, GPT-4) with few-shot prompting. Our results demonstrate that the DeBERTa-based model can master the reasoning task and that the performance of GPTs can improve significantly even when a small number of samples is provided (9 shots). We open-source our code and datasets.
- The Description Logic Handbook: Theory, Implementation, and Applications. Cambridge University Press.
- A multitask, multilingual, multimodal evaluation of chatgpt on reasoning, hallucination, and interactivity.
- Transformers as soft reasoners over language. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI 2020, pages 3882–3890. ijcai.org.
- Explaining answers with entailment trees. In Conference on Empirical Methods in Natural Language Processing.
- BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota. Association for Computational Linguistics.
- Accelerate: Training and inference at scale made simple, efficient and adaptable. https://github.com/huggingface/accelerate.
- Folio: Natural language reasoning with first-order logic.
- Debertav3: Improving deberta using electra-style pre-training with gradient-disentangled embedding sharing.
- Language model analysis for ontology subsumption inference. In Findings of the Association for Computational Linguistics: ACL 2023, pages 3439–3453, Toronto, Canada. Association for Computational Linguistics.
- The manchester OWL syntax. In Proceedings of the OWLED*06 Workshop on OWL: Experiences and Directions, Athens, Georgia, USA, November 10-11, 2006, volume 216 of CEUR Workshop Proceedings. CEUR-WS.org.
- Laconic and precise justifications in OWL. In The Semantic Web - ISWC 2008, 7th International Semantic Web Conference, ISWC 2008, Karlsruhe, Germany, October 26-30, 2008. Proceedings, volume 5318 of Lecture Notes in Computer Science, pages 323–338. Springer.
- The hermit OWL reasoner. In Proceedings of the 1st International Workshop on OWL Reasoner Evaluation (ORE-2012), Manchester, UK, July 1st, 2012, volume 858 of CEUR Workshop Proceedings. CEUR-WS.org.
- Summary of ChatGPT-related research and perspective towards the future of large language models. Meta-Radiology, 1(2):100017.
- Ilya Loshchilov and Frank Hutter. 2019. Decoupled weight decay regularization.
- Logicinference: A new dataset for teaching logical inference to seq2seq models. CoRR, abs/2203.15099.
- OpenAI. 2023. Gpt-4 technical report.
- Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res., 21:140:1–140:67.
- Kyle Richardson and Ashish Sabharwal. 2022. Pushing the limits of rule reasoning in transformers through natural language satisfiability. In Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI 2022, Thirty-Fourth Conference on Innovative Applications of Artificial Intelligence, IAAI 2022, The Twelveth Symposium on Educational Advances in Artificial Intelligence, EAAI 2022 Virtual Event, February 22 - March 1, 2022, pages 11209–11219. AAAI Press.
- PRover: Proof generation for interpretable reasoning over rules. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 122–136, Online. Association for Computational Linguistics.
- Language models are greedy reasoners: A systematic formal analysis of chain-of-thought. CoRR, abs/2210.01240.
- Can transformers reason in fragments of natural language? In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, Abu Dhabi, United Arab Emirates, December 7-11, 2022, pages 11184–11199. Association for Computational Linguistics.
- Proofwriter: Generating implications, proofs, and abductive statements over natural language. In Findings of the Association for Computational Linguistics: ACL/IJCNLP 2021, Online Event, August 1-6, 2021, volume ACL/IJCNLP 2021 of Findings of ACL, pages 3621–3634. Association for Computational Linguistics.
- Large language models are in-context semantic reasoners rather than symbolic reasoners.
- Diagnosing the first-order logical reasoning ability through logicnli. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, Virtual Event / Punta Cana, Dominican Republic, 7-11 November, 2021, pages 3738–3747. Association for Computational Linguistics.
- Stephan Tobies. 2000. The complexity of reasoning with cardinality restrictions and nominals in expressive description logics. J. Artif. Intell. Res., 12:199–217.
- Llama 2: Open foundation and fine-tuned chat models.
- Towards ai-complete question answering: A set of prerequisite toy tasks. In 4th International Conference on Learning Representations, ICLR 2016, San Juan, Puerto Rico, May 2-4, 2016, Conference Track Proceedings.
- Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 38–45, Online. Association for Computational Linguistics.
- Logical reasoning over natural language as knowledge representation: A survey. CoRR, abs/2303.12023.
- On the paradox of learning to reason from data. CoRR, abs/2205.11502.
- Angelos Poulis (2 papers)
- Eleni Tsalapati (2 papers)
- Manolis Koubarakis (15 papers)