Fine-grained explainability for complex inference chains

Achieve fine-grained explainability for complex inference chains in Neuro-Symbolic AI by making each inferential step and its dependence on symbolic rules and neural components interpretable.

Background

Although some systems integrate logic into loss functions or constraints to improve trustworthiness, the authors note that producing detailed, step-level explanations for multi-step inference remains an open problem.

References

Open research questions remain in Neuro-Symbolic AI, including how to develop incremental learning that allows symbolic systems to evolve with new experiences, create context-aware inference mechanisms that adjust reasoning based on situational cues, achieve fine-grained explainability for complex inference chains, and explore meta-cognitive abilities enabling systems to monitor, evaluate, and optimize their learning processes in dynamic environments.

Neuro-Symbolic AI in 2024: A Systematic Review (2501.05435 - Colelough et al., 9 Jan 2025) in Section 4.2 Learning and Inference