Symbol Correctness in Deep Neural Networks Containing Symbolic Layers (2402.03663v1)
Abstract: To handle AI tasks that combine perception and logical reasoning, recent work introduces Neurosymbolic Deep Neural Networks (NS-DNNs), which contain -- in addition to traditional neural layers -- symbolic layers: symbolic expressions (e.g., SAT formulas, logic programs) that are evaluated by symbolic solvers during inference. We identify and formalize an intuitive, high-level principle that can guide the design and analysis of NS-DNNs: symbol correctness, the correctness of the intermediate symbols predicted by the neural layers with respect to a (generally unknown) ground-truth symbolic representation of the input data. We demonstrate that symbol correctness is a necessary property for NS-DNN explainability and transfer learning (despite being in general impossible to train for). Moreover, we show that the framework of symbol correctness provides a precise way to reason and communicate about model behavior at neural-symbolic boundaries, and gives insight into the fundamental tradeoffs faced by NS-DNN training algorithms. In doing so, we both identify significant points of ambiguity in prior work, and provide a framework to support further NS-DNN developments.
- Automatic Differentiation in Machine Learning: a Survey. Journal of Machine Learning Research, 18(153):1–43, 2017.
- From SMT to ASP: Solver-Based Approaches to Solving Datalog Synthesis-as-Rule-Selection Problems. Proceedings of the ACM on Programming Languages, 7(POPL):185–217, 2023. doi: 10.1145/3571200.
- Strictly Declarative Specification of Sophisticated Points-to Analyses. In ACM SIGPLAN Conference on Object-Oriented Programming, Systems, Languages, and Applications, pp. 243–262, 2009. doi: 10.1145/1640089.1640108.
- Answer Set Programming at a Glance. Communications of the ACM, 54(12):92–103, 2011. doi: 10.1145/2043174.2043195.
- What You Always Wanted to Know About Datalog (And Never Dared to Ask). IEEE Transactions on Knowledge and Data Engineering, 1(1):146–166, 1989. doi: 10.1109/69.43410.
- Assessing SATNet’s Ability to Solve the Symbol Grounding Problem. In Advances in Neural Information Processing Systems, 2020.
- Neurosymbolic Programming. Foundations and Trends® in Programming Languages, 7(3):1–86, 2021. doi: 10.1561/2500000049.
- Neurosymbolic AI: the 3rd wave. Artificial Intelligence Review, 56(11):12387–12406, 2023. doi: 10.1007/S10462-023-10448-W.
- Specifying and Reasoning About Dynamic Access-Control Policies. In International Joint Conference on Automated Reasoning, pp. 632–646, 2006. doi: 10.1007/11814771˙51.
- Multi-shot ASP solving with Clingo. Theory and Practice of Logic Programming, 19(1):27–82, 2019. doi: 10.1017/S1471068418000054.
- Harnad, S. The symbol grounding problem. Physica D: Nonlinear Phenomena, 42(1–3):335–346, 1990. doi: 10.1016/0167-2789(90)90087-6.
- Scallop: From Probabilistic Deductive Databases to Scalable Differentiable Reasoning. In Advances in Neural Information Processing Systems, pp. 25134–25145, 2021.
- An Algebraic Prolog for Reasoning about Possible Worlds. In AAAI Conference on Artificial Intelligence, pp. 209–214, 2011.
- Datalog with Constraints: A Foundation for Trust Management Languages. In International Symposium on Practical Aspects of Declarative Languages, pp. 58–73, 2003. doi: 10.1007/3-540-36388-2˙6.
- Closed Loop Neural-Symbolic Learning via Integrating Neural Perception, Grammar Parsing, and Symbolic Reasoning. In International Conference on Machine Learning, pp. 5884–5894, 2020.
- Scallop: A Language for Neurosymbolic Programming. Proceedings of the ACM on Programming Languages, 7(PLDI):1463–1487, 2023a. doi: 10.1145/3591280.
- Softened Symbol Grounding for Neuro-symbolic Systems. In International Conference on Learning Representations, 2023b.
- Declarative Networking: Language, Execution and Optimization. In ACM SIGMOD International Conference on Management of Data, pp. 97–108, 2006. doi: 10.1145/1142473.1142485.
- Neural probabilistic logic programming in DeepProbLog. Artificial Intelligence, 298:103504, 2021. doi: 10.1016/j.artint.2021.103504.
- The Neuro-Symbolic Concept Learner: Interpreting Scenes, Words, and Sentences From Natural Supervision. In International Conference on Learning Representations, 2019.
- PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems, pp. 8024–8035, 2019.
- Techniques for Symbol Grounding with SATNet. In Advances in Neural Information Processing Systems, pp. 20733–20744, 2021.
- Differentiation of Blackbox Combinatorial Solvers. In International Conference on Learning Representations, 2020.
- SATNet: Bridging deep learning and logical reasoning using a differentiable satisfiability solver. In International Conference on Machine Learning, pp. 6545–6554, 2019.
- Grounding Neural Inference with Satisfiability Modulo Theories. In Advances in Neural Information Processing Systems, 2023.
- Cloning-Based Context-Sensitive Pointer Alias Analysis Using Binary Decision Diagrams. In ACM SIGPLAN Conference on Programming Language Design and Implementation, pp. 131–144, 2004. doi: 10.1145/996841.996859.
- NeurASP: Embracing Neural Networks into Answer Set Programming. In International Joint Conference on Artificial Intelligence, pp. 1755–1762, 2020. doi: 10.24963/IJCAI.2020/243.
- Neural-Symbolic VQA: Disentangling Reasoning from Vision and Language Understanding. In Advances in Neural Information Processing Systems, pp. 1039–1050, 2018.