Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
51 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Thought-Like-Pro: Enhancing Reasoning of Large Language Models through Self-Driven Prolog-based Chain-of-Thought (2407.14562v2)

Published 18 Jul 2024 in cs.AI and cs.CL
Thought-Like-Pro: Enhancing Reasoning of Large Language Models through Self-Driven Prolog-based Chain-of-Thought

Abstract: LLMs have shown exceptional performance as general-purpose assistants, excelling across a variety of reasoning tasks. This achievement represents a significant step toward achieving artificial general intelligence (AGI). Despite these advancements, the effectiveness of LLMs often hinges on the specific prompting strategies employed, and there remains a lack of a robust framework to facilitate learning and generalization across diverse reasoning tasks. To address these challenges, we introduce a novel learning framework, THOUGHT-LIKE-PRO In this framework, we utilize imitation learning to imitate the Chain-of-Thought (CoT) process which is verified and translated from reasoning trajectories generated by a symbolic Prolog logic engine. This framework proceeds in a self-driven manner, that enables LLMs to formulate rules and statements from given instructions and leverage the symbolic Prolog engine to derive results. Subsequently, LLMs convert Prolog-derived successive reasoning trajectories into natural language CoT for imitation learning. Our empirical findings indicate that our proposed approach substantially enhances the reasoning abilities of LLMs and demonstrates robust generalization across out-of-distribution reasoning tasks.

Enhancing LLMs' Reasoning Through Prolog-Based Chain-of-Thought

The paper "Thought-Like-Pro: Enhancing Reasoning of LLMs through Self-Driven Prolog-based Chain-of-Thought" introduces a framework aimed at improving the reasoning capabilities of LLMs. The primary innovation lies in the integration of Prolog logic engines to guide the development of coherent and logically sound reasoning trajectories, which LLMs can then learn to imitate. This approach addresses a significant challenge in the current landscape of LLMs—namely, the dependency on specific prompting strategies and the lack of a unified framework for consistent reasoning across diverse tasks.

The proposed framework, Thought-Like-Pro, utilizes imitation learning to capture the Chain-of-Thought (CoT) process informed by the reasoning paths generated by Prolog. This self-driven methodology allows LLMs to construct their rules and perform inference using the Prolog engine, subsequently translating these logic-based pathways into natural language CoT reasoning. The framework's empirical evaluations demonstrate that this method significantly enhances LLMs' reasoning proficiency and generalizes well to out-of-distribution (OOD) reasoning tasks.

In the methods section, the authors describe the progressive steps involved in the Thought-Like-Pro framework. It begins with the formulation of rules and facts by employing open-source chat LLMs like Llama3-8B-Instruct, followed by symbolic reasoning through Prolog, which verifies these paths against known truths. Thus, the validity of the reasoning process is ensured before the natural language translation occurs for imitation learning. This iterative process helps align LLMs' reasoning steps with logical structures, enhancing performance across a broad spectrum of reasoning challenges.

Experimental evaluations conducted on datasets such as GSM8K, ProofWriter, and PrOntoQA reveal a marked improvement in reasoning accuracy when using Thought-Like-Pro. This improvement is attributed to the training on verified logical trajectories, which bolsters the LLMs' capacity to handle both structured logical reasoning tasks and complex algorithmic computations. The paper also explores model averaging techniques to prevent catastrophic forgetting, balancing domain-specific fine-tuning with the retention of general reasoning capabilities.

The authors acknowledge certain limitations in their paper, notably the scale of tasks addressed and the potential variances in OOD generalization. However, they assert that the Thought-Like-Pro framework offers a robust foundation for extending logical reasoning capabilities, especially in industrial applications requiring reliable and explainable AI systems.

The paper’s implications suggest potential pathways for evolving the capabilities of LLMs, pointing towards a future where AI systems can autonomously construct and follow logical reasoning processes with minimal human supervision. The fusion of symbolic logic and data-driven learning models, as demonstrated, may enhance AI's application in domains that demand stringent logical frameworks, such as legal reasoning, complex decision-making, and mathematical problem-solving.

Future developments in AI could further capitalize on this framework to unify various reasoning paradigms, thereby enhancing LLMs' role as general-purpose intelligent agents. This would contribute significantly to the overarching goal of approaching artificial general intelligence (AGI) by fostering more natural and consistent reasoning abilities in machines.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Xiaoyu Tan (21 papers)
  2. Yongxin Deng (6 papers)
  3. Xihe Qiu (14 papers)
  4. Weidi Xu (10 papers)
  5. Chao Qu (39 papers)
  6. Wei Chu (118 papers)
  7. Yinghui Xu (48 papers)
  8. Yuan Qi (85 papers)
Citations (2)