Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Neural-Symbolic Learning and Reasoning: A Survey and Interpretation (1711.03902v1)

Published 10 Nov 2017 in cs.AI

Abstract: The study and understanding of human behaviour is relevant to computer science, artificial intelligence, neural computation, cognitive science, philosophy, psychology, and several other areas. Presupposing cognition as basis of behaviour, among the most prominent tools in the modelling of behaviour are computational-logic systems, connectionist models of cognition, and models of uncertainty. Recent studies in cognitive science, artificial intelligence, and psychology have produced a number of cognitive models of reasoning, learning, and language that are underpinned by computation. In addition, efforts in computer science research have led to the development of cognitive computational systems integrating machine learning and automated reasoning. Such systems have shown promise in a range of applications, including computational biology, fault diagnosis, training and assessment in simulators, and software verification. This joint survey reviews the personal ideas and views of several researchers on neural-symbolic learning and reasoning. The article is organised in three parts: Firstly, we frame the scope and goals of neural-symbolic computation and have a look at the theoretical foundations. We then proceed to describe the realisations of neural-symbolic computation, systems, and applications. Finally we present the challenges facing the area and avenues for further research.

Neural-Symbolic Learning and Reasoning: Comprehensive Overview

The paper "Neural-Symbolic Learning and Reasoning: A Survey and Interpretation" by Besold et al. embarks on a comprehensive examination of the interplay between neural networks and symbolic logic—a frontier in addressing high-level cognitive tasks with neural-symbolic computation. The authors meticulously dissect and contribute to the domain by scrutinizing the essential components, theoretical foundations, systems, applications, and ongoing challenges. This work is pivotal in understanding the collaboration between subsymbolic neural networks and symbolic reasoning paradigms.

The underlying motivation for neural-symbolic computation stems from a significant gap in representing and processing complex combinatorial structures inherent in cognitive reasoning within conventional ANN frameworks. The authors survey the utility of integrating machine learning with rich symbolic representations to encapsulate cognitive systems' multi-faceted nature. They discuss the embedding and processing of logical representations, emphasizing challenges such as the binding problem and first-order logic manipulation in connectionist architectures.

Tackling first-order logic learning, Besold et al. highlight how approximating the consequence operators of logic programs using feed-forward networks and homeomorphic embeddings transcends the bounds of propositional logic. The discourse extends to integrating uncertainty via Markov Logic Networks (MLNs), providing a probabilistic framework leveraging the tractability of graphical models combined with the expressiveness of logic, demonstrating the equivalence of neural-symbolic systems to probabilistic model structures.

Practically, these neural-symbolic systems prove vital across numerous applications like training simulators—where translation from symbolic to neural representations enhances real-time decision-making under uncertainty. Each of these applications underscores the potential of neural-symbolic systems in bridging high-level abstract reasoning with low-level dynamic processing.

The paper also discusses conceptual frameworks like NSCA, showcasing architecture that effectively integrates symbolic knowledge and neural learning. Furthermore, challenges such as scalable, robust extraction of symbolic knowledge from expansive network models and effective real-world application integration are identified, urging for multi-disciplinary collaboration and innovative methodologies.

In conclusion, while remarkable strides have been made in neural-symbolic integration, significant research questions remain. Core among these are the full reconciliation of logic-oriented learning with richly probabilistic neural representations, the development of more scalable algorithms, and the production of comprehensive frameworks unifying these approaches. The substantive discourse within this paper positions neural-symbolic integration as a cornerstone for advancing cognitive sciences, AI, and machine learning, propelling towards accomplishing human-level artificial intelligence.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (14)
  1. Tarek R. Besold (8 papers)
  2. Artur d'Avila Garcez (29 papers)
  3. Sebastian Bader (19 papers)
  4. Howard Bowman (13 papers)
  5. Pedro Domingos (17 papers)
  6. Pascal Hitzler (41 papers)
  7. Kai-Uwe Kuehnberger (2 papers)
  8. Luis C. Lamb (22 papers)
  9. Daniel Lowd (23 papers)
  10. Priscila Machado Vieira Lima (4 papers)
  11. Leo de Penning (1 paper)
  12. Gadi Pinkas (3 papers)
  13. Hoifung Poon (61 papers)
  14. Gerson Zaverucha (2 papers)
Citations (305)
Youtube Logo Streamline Icon: https://streamlinehq.com