A Neuro-Symbolic Approach to Multi-Agent RL for Interpretability and Probabilistic Decision Making (2402.13440v1)
Abstract: Multi-agent reinforcement learning (MARL) is well-suited for runtime decision-making in optimizing the performance of systems where multiple agents coexist and compete for shared resources. However, applying common deep learning-based MARL solutions to real-world problems suffers from issues of interpretability, sample efficiency, partial observability, etc. To address these challenges, we present an event-driven formulation, where decision-making is handled by distributed co-operative MARL agents using neuro-symbolic methods. The recently introduced neuro-symbolic Logical Neural Networks (LNN) framework serves as a function approximator for the RL, to train a rules-based policy that is both logical and interpretable by construction. To enable decision-making under uncertainty and partial observability, we developed a novel probabilistic neuro-symbolic framework, Probabilistic Logical Neural Networks (PLNN), which combines the capabilities of logical reasoning with probabilistic graphical models. In PLNN, the upward/downward inference strategy, inherited from LNN, is coupled with belief bounds by setting the activation function for the logical operator associated with each neural network node to a probability-respecting generalization of the Fr\'echet inequalities. These PLNN nodes form the unifying element that combines probabilistic logic and Bayes Nets, permitting inference for variables with unobserved states. We demonstrate our contributions by addressing key MARL challenges for power sharing in a system-on-chip application.
- Cozman, F. G. 2000. Credal networks. Artificial Intelligence, 120(2): 199–233.
- Foundations of Reasoning with Uncertainty via Real-valued Logics. CoRR, abs/2008.02429.
- Fréchet, M. 1934. Généralisations du theéorème des probabilités totales. Fundamenta Mathematicae, 25: 379–387.
- Gottwald, S. 2022. Many-Valued Logic. In Zalta, E. N., ed., The Stanford Encyclopedia of Philosophy. Metaphysics Research Lab, Stanford University, Summer 2022 edition.
- Planning and acting in partially observable stochastic domains. Artificial intelligence, 101(1-2): 99–134.
- Logical Credal Networks. In Thirty-sixth Conference on Neural Information Processing Systems (NeurIPS).
- Thirty years of credal networks: Specification, algorithms and complexity. International Journal of Approximate Reasoning, 126: 133–157.
- Efficient Offline Communication Policies for Factored Multiagent POMDPs. In Shawe-Taylor, J.; Zemel, R.; Bartlett, P.; Pereira, F.; and Weinberger, K., eds., Advances in Neural Information Processing Systems, volume 24. Curran Associates, Inc.
- Messias, J. V. T. 2014. Decision-making under uncertainty for real robot teams. Ph.D. thesis, Instituto Superior Técnico.
- Pearl, J. 1988. Probabilistic Reasoning in Intelligent Systems. Morgan Kaufmann.
- Logical Neural Networks. arXiv:2006.13155.
- Logical Neural Networks for Knowledge Base Completion with Embeddings & Rules. In Goldberg, Y.; Kozareva, Z.; and Zhang, Y., eds., Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, Abu Dhabi, United Arab Emirates, December 7-11, 2022, 3863–3875. Association for Computational Linguistics.
- Neuro-Symbolic Inductive Logic Programming with Logical Neural Networks. In Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI 2022, Thirty-Fourth Conference on Innovative Applications of Artificial Intelligence, IAAI 2022, The Twelveth Symposium on Educational Advances in Artificial Intelligence, EAAI 2022 Virtual Event, February 22 - March 1, 2022, 8212–8219. AAAI Press.
- Policy gradient methods for reinforcement learning with function approximation. Advances in neural information processing systems, 12.