Introduction to Neurosymbolic Reasoning with LLMs
Symbolic reasoning is a critical component of many applications, requiring the capability to handle tasks that involve logical operations, sequential planning, or common-sense reasoning. Recent advancements in AI have seen LLMs like GPT-4 significantly improve performance on various reasoning tasks. This brings to light their potential application as symbolic reasoners. One area where this potential is being investigated is text-based games, which serve as robust benchmarks for language agents to exhibit symbolic reasoning capabilities.
Agent Design and Symbolic Modules
In order to harness the power of LLMs for symbolic reasoning, a novel agent is proposed to interact with text-based game environments and accompanying symbolic modules. The paper presents a methodology where the agent commences with a role initialization, then processes observations and a set of valid actions provided by the game environment along with an external symbolic module. Based on these inputs, the agent selects an appropriate action in a zero-shot manner—without additional training. Symbolic modules are specialized tools, such as calculators or navigators, that enhance the agent's reasoning capabilities and broaden the spectrum of tasks LLMs can undertake.
Experimental Results and Benchmarking
The performance of the proposed LLM-based agent has been meticulously evaluated across four different text-based games that require symbolic reasoning. Detailed experiments demonstrate that this new approach can substantially outperform existing benchmarks, achieving an impressive average performance of 88% in various tasks. Furthermore, the LLM-based agent manages to achieve this high level of performance without relying on extensive pre-trained data, unlike other deep learning models, which often necessitate a significant amount of expert data for training.
Impact and Future Work
The findings illustrate the substantial promise of LLMs as neurosymbolic reasoners capable of tackling complex symbolic tasks. Not only do these agents surpass traditional models that rely on deep learning and large datasets, but they also offer cost-effective and efficient problem-solving strategies. There remain areas for enhancement, such as improving memory handling in tasks that involve sorting logic. Future research should focus on refining and translating these agents' capabilities to more complex and varied real-world applications, necessitating the integration of more advanced symbolic modules.