Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Transformers in the Service of Description Logic-based Contexts (2311.08941v3)

Published 15 Nov 2023 in cs.CL and cs.AI

Abstract: Recent advancements in transformer-based models have initiated research interests in investigating their ability to learn to perform reasoning tasks. However, most of the contexts used for this purpose are in practice very simple: generated from short (fragments of) first-order logic sentences with only a few logical operators and quantifiers. In this work, we construct the natural language dataset, DELTA$_D$, using the description logic language $\mathcal{ALCQ}$. DELTA$_D$ contains 384K examples, and increases in two dimensions: i) reasoning depth, and ii) linguistic complexity. In this way, we systematically investigate the reasoning ability of a supervised fine-tuned DeBERTa-based model and of two LLMs (GPT-3.5, GPT-4) with few-shot prompting. Our results demonstrate that the DeBERTa-based model can master the reasoning task and that the performance of GPTs can improve significantly even when a small number of samples is provided (9 shots). We open-source our code and datasets.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (30)
  1. The Description Logic Handbook: Theory, Implementation, and Applications. Cambridge University Press.
  2. A multitask, multilingual, multimodal evaluation of chatgpt on reasoning, hallucination, and interactivity.
  3. Transformers as soft reasoners over language. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI 2020, pages 3882–3890. ijcai.org.
  4. Explaining answers with entailment trees. In Conference on Empirical Methods in Natural Language Processing.
  5. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota. Association for Computational Linguistics.
  6. Accelerate: Training and inference at scale made simple, efficient and adaptable. https://github.com/huggingface/accelerate.
  7. Folio: Natural language reasoning with first-order logic.
  8. Debertav3: Improving deberta using electra-style pre-training with gradient-disentangled embedding sharing.
  9. Language model analysis for ontology subsumption inference. In Findings of the Association for Computational Linguistics: ACL 2023, pages 3439–3453, Toronto, Canada. Association for Computational Linguistics.
  10. The manchester OWL syntax. In Proceedings of the OWLED*06 Workshop on OWL: Experiences and Directions, Athens, Georgia, USA, November 10-11, 2006, volume 216 of CEUR Workshop Proceedings. CEUR-WS.org.
  11. Laconic and precise justifications in OWL. In The Semantic Web - ISWC 2008, 7th International Semantic Web Conference, ISWC 2008, Karlsruhe, Germany, October 26-30, 2008. Proceedings, volume 5318 of Lecture Notes in Computer Science, pages 323–338. Springer.
  12. The hermit OWL reasoner. In Proceedings of the 1st International Workshop on OWL Reasoner Evaluation (ORE-2012), Manchester, UK, July 1st, 2012, volume 858 of CEUR Workshop Proceedings. CEUR-WS.org.
  13. Summary of ChatGPT-related research and perspective towards the future of large language models. Meta-Radiology, 1(2):100017.
  14. Ilya Loshchilov and Frank Hutter. 2019. Decoupled weight decay regularization.
  15. Logicinference: A new dataset for teaching logical inference to seq2seq models. CoRR, abs/2203.15099.
  16. OpenAI. 2023. Gpt-4 technical report.
  17. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res., 21:140:1–140:67.
  18. Kyle Richardson and Ashish Sabharwal. 2022. Pushing the limits of rule reasoning in transformers through natural language satisfiability. In Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI 2022, Thirty-Fourth Conference on Innovative Applications of Artificial Intelligence, IAAI 2022, The Twelveth Symposium on Educational Advances in Artificial Intelligence, EAAI 2022 Virtual Event, February 22 - March 1, 2022, pages 11209–11219. AAAI Press.
  19. PRover: Proof generation for interpretable reasoning over rules. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 122–136, Online. Association for Computational Linguistics.
  20. Language models are greedy reasoners: A systematic formal analysis of chain-of-thought. CoRR, abs/2210.01240.
  21. Can transformers reason in fragments of natural language? In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, Abu Dhabi, United Arab Emirates, December 7-11, 2022, pages 11184–11199. Association for Computational Linguistics.
  22. Proofwriter: Generating implications, proofs, and abductive statements over natural language. In Findings of the Association for Computational Linguistics: ACL/IJCNLP 2021, Online Event, August 1-6, 2021, volume ACL/IJCNLP 2021 of Findings of ACL, pages 3621–3634. Association for Computational Linguistics.
  23. Large language models are in-context semantic reasoners rather than symbolic reasoners.
  24. Diagnosing the first-order logical reasoning ability through logicnli. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, Virtual Event / Punta Cana, Dominican Republic, 7-11 November, 2021, pages 3738–3747. Association for Computational Linguistics.
  25. Stephan Tobies. 2000. The complexity of reasoning with cardinality restrictions and nominals in expressive description logics. J. Artif. Intell. Res., 12:199–217.
  26. Llama 2: Open foundation and fine-tuned chat models.
  27. Towards ai-complete question answering: A set of prerequisite toy tasks. In 4th International Conference on Learning Representations, ICLR 2016, San Juan, Puerto Rico, May 2-4, 2016, Conference Track Proceedings.
  28. Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 38–45, Online. Association for Computational Linguistics.
  29. Logical reasoning over natural language as knowledge representation: A survey. CoRR, abs/2303.12023.
  30. On the paradox of learning to reason from data. CoRR, abs/2205.11502.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Angelos Poulis (2 papers)
  2. Eleni Tsalapati (2 papers)
  3. Manolis Koubarakis (15 papers)
X Twitter Logo Streamline Icon: https://streamlinehq.com