Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
101 tokens/sec
Gemini 2.5 Pro Premium
50 tokens/sec
GPT-5 Medium
28 tokens/sec
GPT-5 High Premium
27 tokens/sec
GPT-4o
101 tokens/sec
DeepSeek R1 via Azure Premium
90 tokens/sec
GPT OSS 120B via Groq Premium
515 tokens/sec
Kimi K2 via Groq Premium
220 tokens/sec
2000 character limit reached

SymbolicAI: A framework for logic-based approaches combining generative models and solvers (2402.00854v4)

Published 1 Feb 2024 in cs.LG, cs.AI, cs.SC, and cs.SE

Abstract: We introduce SymbolicAI, a versatile and modular framework employing a logic-based approach to concept learning and flow management in generative processes. SymbolicAI enables the seamless integration of generative models with a diverse range of solvers by treating LLMs as semantic parsers that execute tasks based on both natural and formal language instructions, thus bridging the gap between symbolic reasoning and generative AI. We leverage probabilistic programming principles to tackle complex tasks, and utilize differentiable and classical programming paradigms with their respective strengths. The framework introduces a set of polymorphic, compositional, and self-referential operations for multi-modal data that connects multi-step generative processes and aligns their outputs with user objectives in complex workflows. As a result, we can transition between the capabilities of various foundation models with in-context learning capabilities and specialized, fine-tuned models or solvers proficient in addressing specific problems. Through these operations based on in-context learning our framework enables the creation and evaluation of explainable computational graphs. Finally, we introduce a quality measure and its empirical score for evaluating these computational graphs, and propose a benchmark that compares various state-of-the-art LLMs across a set of complex workflows. We refer to the empirical score as the "Vector Embedding for Relational Trajectory Evaluation through Cross-similarity", or VERTEX score for short. The framework codebase and benchmark are linked below.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (177)
  1. The history began from alexnet: A comprehensive survey on deep learning approaches. arXiv preprint arXiv:1803.01164, 2018.
  2. Voyager – an innovative computational resource for artificial intelligence & machine learning applications in science and engineering. In Practice and Experience in Advanced Research Computing, PEARC ’23’, pp.  278–282, New York, NY, USA, 2023. Association for Computing Machinery. ISBN 9781450399852. doi: 10.1145/3569951.3597597.
  3. J. Andreas. Language models as agent models. CoRR, abs/2212.01681, 2022. doi: 10.48550/arXiv.2212.01681.
  4. RUDDER: return decomposition for delayed rewards. In Advances in Neural Information Processing Systems 32, pp.  13566–13577, 2019.
  5. L. M. Augusto. Computational Logic. Vol. 1: Classical Deductive Computing with Classical Logic. College Publications, London, 2 edition, 2022.
  6. F. Badita. 1337 Use Cases for ChatGPT & other Chatbots in the AI-Driven Era. Google Docs, 2022.
  7. D.M. Beazley. Python Essential Reference. Developer’s library : essential references for programming professionals. Addison-Wesley, 2009. ISBN 9780672329784. URL https://books.google.ro/books?id=Chr1NDlUcI8C.
  8. Neural-symbolic learning and reasoning: A survey and interpretation, 2017.
  9. Graph of thoughts: Solving elaborate problems with large language models. arXiv preprint arXiv:2308.09687, 2023.
  10. Pythia: A suite for analyzing large language models across training and scaling, 2023.
  11. PENG Bo. Blinkdl/rwkv-lm: 0.01. Technical report, Zenodo, August 2021. URL https://doi.org/10.5281/zenodo.5196577.
  12. W. Bradley Knox and Peter Stone. TAMER: Training an Agent Manually via Evaluative Reinforcement. In 2008 7th IEEE International Conference on Development and Learning, pp.  292–297, Monterey, CA, August 2008. IEEE. ISBN 978-1-4244-2661-4. doi: 10.1109/DEVLRN.2008.4640845.
  13. Contextual prerequisites for understanding: Some investigations of comprehension and recall. Journal of Verbal Learning and Verbal Behavior, 11(6):717–726, 1972. ISSN 0022-5371.
  14. Language models are few-shot learners. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin (eds.), Advances in Neural Information Processing Systems, volume 33, pp.  1877–1901. Curran Associates, Inc., 2020.
  15. H. Chase. LangChain. Technical report, LangChain, 01 2023. URL https://github.com/hwchase17/langchain.
  16. Hilog: A foundation for higher-order logic programming. The Journal of Logic Programming, 15(3):187–230, 1993. ISSN 0743-1066.
  17. N. Chomsky. Three models for the description of language. IRE Transactions on Information Theory, 2(3):113–124, 1956. doi: 10.1109/TIT.1956.1056813.
  18. Deep reinforcement learning from human preferences. Advances in neural information processing systems, 30, 2017.
  19. A. Colmerauer and P. Roussel. The birth of Prolog. In HOPL-II, 1993.
  20. Neurosymbolic ai: The 3rd wave. arXiv preprint arXiv:2012.05876, 2020.
  21. Neural-symbolic computing: An effective methodology for principled integration of machine learning and reasoning. Journal of Applied Logic, 2019.
  22. A. Dawid and Y. LeCun. Introduction to latent variable energy-based models: A path towards autonomous machine intelligence. arXiv preprint arXiv:2306.02572, 2023.
  23. J. Degrave. Building A Virtual Machine inside ChatGPT. Technical report, Engraved, 11 2022. URL https://www.engraved.blog/building-a-virtual-machine-inside/.
  24. The representation of semantic information across human cerebral cortex during listening versus reading is invariant to stimulus modality. Journal of Neuroscience, 39(39):7722–7736, 2019. ISSN 0270-6474. doi: 10.1523/JNEUROSCI.0675-19.2019.
  25. D. C. Dennett. Real patterns. Journal of Philosophy, 88(1):27–51, 1991. doi: 10.2307/2027085.
  26. Understanding software-2.0: A study of machine learning library usage and evolution. ACM Transactions on Software Engineering and Methodology (TOSEM), 30(4):55:1–55:42, jul 2021. ISSN 1049-331X. doi: 10.1145/3453478.
  27. Xai and strategy extraction via reward redistribution. In A. Holzinger, R. Goebel, R. Fong, T. Moon, K.-R. Müller, and W. Samek (eds.), xxAI - Beyond Explainable AI: International Workshop, Held in Conjunction with ICML 2020, July 18, 2020, Vienna, Austria, Revised and Extended Papers, pp.  177–205, Cham, 2022. Springer International Publishing. ISBN 978-3-031-04083-2. doi: 10.1007/978-3-031-04083-2˙10.
  28. Addressing parameter choice issues in unsupervised domain adaptation by aggregation. In The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023. OpenReview.net, 2023.
  29. Logic tensor networks for semantic image interpretation. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, IJCAI-17, pp.  1596–1602, 2017.
  30. The fréchet distance between multivariate normal distributions. Journal of Multivariate Analysis, 12(3):450–455, 1982. doi: https://doi.org/10.1016/0047-259X(82)90077-X.
  31. Kevin Ellis. Human-like few-shot learning via bayesian reasoning over natural language. arXiv preprint arXiv:2306.02797, 2023.
  32. Large language models are neurosymbolic reasoners. arXiv preprint arXiv:2401.09334, 2024.
  33. New method for fMRI investigations of language: Defining rois functionally in individual subjects. Journal of neurophysiology, 104:1177–94, 08 2010. doi: 10.1152/jn.00032.2010.
  34. Dendral, 1965.
  35. Promptbreeder: Self-referential self-improvement via prompt evolution. arXiv preprint arXiv:2309.16797, 2023.
  36. Applying formal verification methods to rule-based programs. Int. J. Expert Syst., 7(3):203–237, sep 1994. ISSN 0894-9077.
  37. S. Ganguly and V. Pudi. Paper2vec: Combining graph and text information for scientific paper representation. In Joemon Jose et al. (eds.), Advances in Information Retrieval, volume 10193 of Lecture Notes in Computer Science. Springer, Cham, 2017. ISBN 978-3-319-56607-8. doi: 10.1007/978-3-319-56608-5˙30.
  38. Retrieval-augmented generation for large language models: A survey. arXiv preprint arXiv:2312.10997, 2023.
  39. Neural-symbolic learning and reasoning: Contributions and challenges. In AAAI Conference, 2015.
  40. Few-Shot Learning by Dimensionality Reduction in Gradient Space. arXiv preprint arXiv:2206.03483, 2022.
  41. Koala: A dialogue model for academic research. Blog post, April 2023. URL https://bair.berkeley.edu/blog/2023/04/03/koala/.
  42. Google. Gemini: A family of highly capable multimodal models. arXiv preprint arXiv:2312.11805, 2023.
  43. Retrieval-augmented reinforcement learning. In K. Chaudhuri, S. Jegelka, L. Song, C. Szepesvari, G. Niu, and S. Sabato (eds.), Proceedings of the 39th International Conference on Machine Learning, volume 162 of Proceedings of Machine Learning Research, pp.  7740–7765. PMLR, 17–23 Jul 2022.
  44. A kernel two-sample test. Journal of Machine Learning Research, 13(25):723–773, 2012.
  45. A. Gu and T. Dao. Mamba: Linear-time sequence modeling with selective state spaces. arXiv preprint arXiv:2312.00752, 2023.
  46. CycleGT: Unsupervised graph-to-text and text-to-graph generation via cycle training. arXiv preprint arXiv:2006.04702, 2020.
  47. Is neuro-symbolic AI meeting its promises in natural language processing? a structured review. Semantic Web, pp.  1–42, nov 2022. doi: 10.3233/sw-223228.
  48. Array programming with NumPy. Nature, 585(7825):357–362, 2020. doi: 10.1038/s41586-020-2649-2.
  49. Gans trained by a two time-scale update rule converge to a local nash equilibrium. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS’17, pp.  6629–6640, Red Hook, NY, USA, 2017. Curran Associates Inc. ISBN 9781510860964.
  50. S. Hochreiter. Toward a broad ai. Commun. ACM, 65(4):56–57, mar 2022. ISSN 0001-0782.
  51. S. Hochreiter and J. Schmidhuber. Flat minima. Neural Comput., 9(1):1–42, 1997.
  52. Convergence proof for actor-critic methods applied to PPO and RUDDER. arXiv preprint arXiv:2012.01399, 2020.
  53. Multilayer feedforward networks are universal approximators. Neural Networks, 2:359–366, 1989. doi: 10.1016/0893-6080(89)90020-8.
  54. Precision fMRI reveals that the language-selective network supports both phrase-structure building and lexical access during language production. bioRxiv, 2022. doi: 10.1101/2021.09.10.459596.
  55. Harnessing deep neural networks with logic rules. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp.  2410–2420, Berlin, Germany, August 2016. Association for Computational Linguistics.
  56. Natural speech reveals the semantic maps that tile human cerebral cortex. Nature, 532(7600):453–458, 2016. doi: 10.1038/nature17637.
  57. Reward learning from human preferences and demonstrations in atari. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett (eds.), Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018.
  58. A generalist neural algorithmic learner. In LOG IN, 2022.
  59. Wolfram Research, Inc. Mathematica, Version 13.2, 2022. URL https://www.wolfram.com/mathematica. Champaign, IL.
  60. Neuromorphic silicon neuron circuits. Frontiers in Neuroscience, 5, 2011. ISSN 1662-453X. doi: 10.3389/fnins.2011.00073.
  61. Jigsaw: Large language models meet program synthesis. arXiv, 2021.
  62. Billion-scale similarity search with GPUs. IEEE Transactions on Big Data, 7(3):535–547, 2019.
  63. E. Jones and J. Steinhardt. Capturing failures of large language models via human cognitive biases. arXiv preprint arXiv:2202.12299, 2022.
  64. A. Karpathy. Software 2.0. Medium, 2017. URL https://karpathy.medium.com/software-2-0-a64152b37c35.
  65. Are Pretrained Language Models Symbolic Reasoners over Knowledge? In R. Fernández and T. Linzen (eds.), Proceedings of the 24th Conference on Computational Natural Language Learning, CoNLL 2020, Online, November 19-20, 2020, pp.  552–564. Association for Computational Linguistics, 2020. doi: 10.18653/v1/2020.conll-1.45.
  66. T. Katsch. Gateloop: Fully data-controlled linear recurrence for sequence modeling. arXiv preprint arXiv:2311.01927, 2023.
  67. Toward trustworthy neural program synthesis. arXiv preprint arXiv:2210.00848, 2023.
  68. Language models can solve computer tasks, 2023.
  69. L. Kirsch and J. Schmidhuber. Eliminating meta optimization through self-referential meta learning. arXiv preprint arXiv:2212.14392, 2022.
  70. Openassistant conversations – democratizing large language model alignment, 2023.
  71. Hybrid systems: the equivalence of rule-based expert system and artificial neural network inference. Int. J. Expert Syst., 4(3):281–297, jan 1991. ISSN 0894-9077.
  72. Causal Reasoning and Large Language Models: Opening a New Frontier for Causality. arXiv, 2023.
  73. J. E. Laird. Introduction to soar, 2022.
  74. Soar: An architecture for general intelligence. Artificial Intelligence, 33(1):1–64, 1987. ISSN 0004-3702.
  75. Graph neural networks meet neural-symbolic computing: A survey and perspective. In AAAI Conference, 2020.
  76. Cognitive architectures: Research issues and challenges. Cognitive Systems Research, 10:141–160, 2009. doi: 10.1016/j.cogsys.2006.07.004.
  77. A survey on retrieval-augmented text generation. arXiv preprint arXiv:2202.01110, 2022a.
  78. Competition-level code generation with alphacode. Science, 378(6624):1092–1097, 2022b.
  79. Reinforcement learning with human feedback: Learning dynamic choices via pessimism. arXiv preprint arxiv:2305.18438, 2023.
  80. Lost in the middle: How language models use long contexts. arXiv preprint arXiv:2307.03172, 2023.
  81. Codexglue: A machine learning benchmark dataset for code understanding and generation. arXiv preprint arXiv:2102.04664, 2021.
  82. M. Lutz. Learning Python: Powerful Object-Oriented Programming. Animal Guide. O’Reilly Media, 2013. ISBN 9781449355715.
  83. Faithful chain-of-thought reasoning, 2023.
  84. Interactive Learning from Policy-Dependent Human Feedback. In Proceedings of the 34th International Conference on Machine Learning, pp.  2285–2294. PMLR, July 2017.
  85. M. Macsweeney. Neural systems underlying british sign language and audio-visual english processing in native users. Brain, 125:1583–1593, 07 2002. doi: 10.1093/brain/awf153.
  86. Self-refine: Iterative refinement with self-feedback, 2023.
  87. Dissociating language and thought in large language models: a cognitive perspective. CoRR, abs/2301.06627, 2023.
  88. DeepProbLog: Neural Probabilistic Logic Programming. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett (eds.), Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018.
  89. The neuro-symbolic concept learner: Interpreting scenes, words, and sentences from natural supervision. In 7th International Conference on Learning Representations, ICLR 2019, 2019.
  90. G. Marcus. The Next Decade in AI: Four Steps Towards Robust Artificial Intelligence. arXiv preprint arXiv:2002.06177, 2020.
  91. Python Cookbook. O’Reilly Media, 2005. ISBN 9780596554743. URL https://books.google.ro/books?id=Q0s6Vgb98CQC.
  92. J. McCarthy. Lisp: A programming system for symbolic manipulations. In Preprints of Papers Presented at the 14th National Meeting of the Association for Computing Machinery, ACM ’59, pp.  1–4, New York, NY, USA, 1959. Association for Computing Machinery. ISBN 9781450373647. doi: 10.1145/612201.612243.
  93. A proposal for the dartmouth summer research project on artificial intelligence, august 31, 1955. AI magazine, 27(4):12–12, 2006.
  94. W. S. McCulloch and W. Pitts. A Logical Calculus of Ideas Immanent in Nervous Activity. Bulletin of Mathematical Biophysics, 5:115–133, 1943. doi: 10.1007/BF02478255.
  95. Shared language: Overlap and segregation of the neuronal infrastructure for speaking and listening revealed by functional mri. Psychological Science, 22(9):1173–1182, 2011. doi: 10.1177/0956797611418347. PMID: 21841148.
  96. Microsoft. Bing is your AI-powered copilot for the web. Technical report, Microsoft, 2023. URL https://bing.com/chat.
  97. Efficient Estimation of Word Representations in Vector Space. International Conference on Learning Representations, 2013.
  98. Retrospective analysis of the 2019 minerl competition on sample efficient reinforcement learning. In H. J. Escalante and R. Hadsell (eds.), Proceedings of the NeurIPS 2019 Competition and Demonstration Track, volume 123 of Proceedings of Machine Learning Research, pp.  203–214. PMLR, Dec 2020.
  99. L. De Moura and N. Bjørner. Z3: an efficient smt solver. In Proceedings of the Theory and Practice of Software, 14th International Conference on Tools and Algorithms for the Construction and Analysis of Systems, TACAS’08/ETAPS’08, pp.  337–340, Berlin, Heidelberg, 2008. Springer-Verlag. ISBN 3540787992.
  100. A. Newell. Unified Theories of Cognition. Harvard University Press, USA, 1990. ISBN 0674920996.
  101. A. Newell and H. Simon. The logic theory machine–a complex information processing system. IRE Transactions on information theory, 2(3):61–79, 1956.
  102. A. Newell and H. A. Simon. Human problem solving. Prentice-Hall, pp.  920, 1972.
  103. A. Newell and H. A. Simon. Computer science as empirical inquiry: symbols and search. Commun. ACM, 19(3):113–126, mar 1976. ISSN 0001-0782. doi: 10.1145/360018.360022.
  104. Empirical explorations of the logic theory machine: a case study in heuristic. IRE-AIEE-ACM ’57 (Western): Papers presented at the February 26-28, 1957, western joint computer conference: Techniques for reliability, pp.  218–230, 1957. doi: 10.1145/1455567.1455605.
  105. Can generalist foundation models outcompete special-purpose tuning? case study in medicine. arXiv preprint arXiv:2311.16452, 2023.
  106. Wavenet: A generative model for raw audio. arXiv preprint arXiv:1609.03499, 2016.
  107. OpenAI. Introducing ChatGPT. Technical report, OpenAI, November 2022. URL https://openai.com/blog/chatgpt.
  108. OpenAI. GPT-4 Technical Report. arXiv, 2023.
  109. Training language models to follow instructions with human feedback. arXiv preprint arXiv:2203.02155, 2022.
  110. History compression via language models in reinforcement learning. In K. Chaudhuri, S. Jegelka, L. Song, C. Szepesvari, G. Niu, and S. Sabato (eds.), Proceedings of the 39th International Conference on Machine Learning, volume 162 of Proceedings of Machine Learning Research, pp.  17156–17185. PMLR, July 2022.
  111. Semantic helm: An interpretable memory for reinforcement learning. CoRR, abs/2306.09312, 2023. doi: 10.48550/arXiv.2306.09312.
  112. Bridging the domain gap by clustering-based image-text graph matching. arXiv preprint arXiv:2310.02692, 2023.
  113. Pytorch: An imperative style, high-performance deep learning library. arXiv preprint arXiv:1912.01703, abs/1912.01703, 2019.
  114. Align-RUDDER: Learning from few demonstrations by reward redistribution. arXiv preprint arXiv:2009.14108, 2020.
  115. Language Models as Knowledge Bases? In K. Inui, J. Jiang, V. Ng, and X. Wan (eds.), Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP 2019, Hong Kong, China, November 3-7, 2019, pp.  2463–2473. Association for Computational Linguistics, 2019. doi: 10.18653/v1/D19-1250.
  116. Boosted prompt ensembles for large language models, 2023.
  117. Hyena hierarchy: Towards larger convolutional language models. arXiv preprint arXiv:2302.10866, 2023.
  118. Dreamfusion: Text-to-3d using 2d diffusion. arXiv preprint arXiv:2209.14988, 2022.
  119. M. Qu and J. Tang. Probabilistic logic neural networks for reasoning. In Proceedings of the 33rd International Conference on Neural Information Processing Systems, 2019.
  120. Direct preference optimization: Your language model is secretly a reward model. arXiv preprint arXiv:2305.18290, 2023.
  121. Zero-shot text-to-image generation. In International Conference on Machine Learning, pp.  8821–8831. PMLR, 2021.
  122. Selective and invariant neural responses to spoken and written narratives. Journal of Neuroscience, 33(40):15978–15988, 2013. ISSN 0270-6474. doi: 10.1523/JNEUROSCI.1580-13.2013.
  123. ReplikaAI. Pushing the Boundaries of AI to Talk to the Dead. Technical report, ReplikaAI, 2016. URL https://www.bloomberg.com/news/articles/2016-10-20/pushing-the-boundaries-of-ai-to-talk-to-the-dead.
  124. Mathematical discoveries from program search with large language models. Nature, 2023. doi: 10.1038/s41586-023-06924-6.
  125. F. Rosenblatt. The perceptron: A probabilistic model for information storage and organization in the brain. Psychological Review, 65(6):386–408, 1958. doi: 10.1037/h0042519.
  126. Large language models are not zero-shot communicators. CoRR, abs/2210.14986, 2022. doi: 10.48550/arXiv.2210.14986.
  127. Learning representations by back-propagating errors. Nature, 323:533–536, 1986. doi: 10.1038/323533a0.
  128. Photorealistic text-to-image diffusion models with deep language understanding. arXiv preprint arXiv:2205.11487, 2022.
  129. Symbolic behaviour in artificial intelligence. arXiv preprint arXiv:2102.03406, 2022.
  130. Toolformer: Language models can teach themselves to use tools, 2023.
  131. J. Schmidhuber. Gödel machines: Fully self-referential optimal universal self-improvers. Cognitive Technologies, 8:199–226, 01 2007. doi: 10.1007/978-3-540-68677-4˙7.
  132. J. Schmidhuber. Driven by compression progress: A simple principle explains essential aspects of subjective beauty, novelty, surprise, interestingness, attention, curiosity, creativity, art, science, music, jokes. arXiv preprint arXiv:0812.4360, 2009.
  133. Mastering atari, go, chess and shogi by planning with a learned model. Nature, 588(7839):604–609, 2020. doi: 10.1038/s41586-020-03051-4.
  134. A dataset perspective on offline reinforcement learning. In Conference on Lifelong Learning Agents, pp.  470–517. PMLR, 2022.
  135. A new fun and robust version of an fMRI localizer for the frontotemporal language system. Cognitive Neuroscience, 8:1–10, 07 2016. doi: 10.1080/17588928.2016.1201466.
  136. Reflexion: an autonomous agent with dynamic memory and self-reflection, 2023.
  137. Mastering the game of Go with deep neural networks and tree search. Nature, 529(7587):484–489, 2016. doi: 10.1038/nature16961.
  138. Mastering chess and shogi by self-play with a general reinforcement learning algorithm. arXiv preprint arXiv:1712.01815, 2017a.
  139. Mastering the game of go without human knowledge. Nature, 550:354–359, 2017b. doi: 10.1038/nature24270.
  140. Make-a-video: Text-to-video generation without text-video data. arXiv preprint arXiv:2209.14792, 2022.
  141. Large language models encode clinical knowledge. Nature, 620(7972):172–180, 2023.
  142. Mpnet: Masked and permuted pre-training for language understanding. In Proceedings of the 34th International Conference on Neural Information Processing Systems, NIPS’20, pp.  1414, Red Hook, NY, USA, 2020. Curran Associates Inc. ISBN 9781713829546.
  143. Spotify. Approximate Nearest Neighbors Oh Yeah. Technical report, Spotify, 2017.
  144. Cognitive architectures for language agents. arXiv preprint arXiv:2309.02427, 2023.
  145. M. Summerfield. Programming in Python 3: A Complete Introduction to the Python Language. Developer’s library. Addison-Wesley, 2010. ISBN 9780321680563.
  146. Neurosymbolic programming for science. arXiv preprint arXiv:2210.05050, 2022.
  147. R. S. Sutton. Temporal Credit Assignment in Reinforcement Learning. PhD thesis, University of Massachusetts, Dept. of Comp. and Inf. Sci., 1984.
  148. Comparing model-free and model-based algorithms for offline reinforcement learning. arXiv preprint arXiv:2201.05433, 2022.
  149. Learning theory for distribution regression. J. Mach. Learn. Res., 17(1):5272–5311, Jan 2016. ISSN 1532-4435.
  150. Stanford alpaca: An instruction-following llama model. https://github.com/tatsu-lab/stanford_alpaca, 2023.
  151. Llama: Open and efficient foundation language models, 2023.
  152. Attention is all you need. In I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (eds.), Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017.
  153. P. Veličković and C. Blundell. Neural algorithmic reasoning. Patterns, 2(7):100273, 2021. ISSN 2666-3899. doi: https://doi.org/10.1016/j.patter.2021.100273.
  154. Grammar prompting for domain-specific language generation with large language models. arXiv preprint arXiv:2305.19234, 2023a.
  155. Milvus: A purpose-built vector data management system. In Proceedings of the 2021 International Conference on Management of Data, pp.  2614–2627, 2021a.
  156. Self-consistency improves chain of thought reasoning in language models. arXiv preprint arXiv:2203.11171, 2023b.
  157. Tacotron: Towards end-to-end speech synthesis. arXiv preprint arXiv:1703.10135, 2017.
  158. Codet5: Identifier-aware unified pre-trained encoder-decoder models for code understanding and generation. arXiv preprint arXiv:2109.00859, 2021b.
  159. Self-instruct: Aligning language model with self generated instructions. arXiv preprint arXiv:2212.10560, 2022.
  160. Emergent abilities of large language models. arXiv preprint arXiv:2206.07682, 2022a.
  161. Chain of thought prompting elicits reasoning in large language models. In A. H. Oh, A. Agarwal, D. Belgrave, and K. Cho (eds.), Advances in Neural Information Processing Systems, 2022b.
  162. Large language models are better reasoners with self-verification. arXiv preprint arXiv:2212.09561, 2023.
  163. A. N. Whitehead and B. Russell. Principia Mathematica. Cambridge University Press, 1925–1927.
  164. Modern Hopfield Networks for Return Decomposition for Delayed Rewards. In Deep RL Workshop NeurIPS 2021, 2021.
  165. Trusted artificial intelligence: Towards certification of machine learning applications. arXiv preprint arXiv:2103.16910, 2021.
  166. Writesonic. ChatGPT Alternative Built With Superpowers - ChatSonic. Technical report, Chatsonic, 2022. URL https://writesonic.com/chat.
  167. Baize: An open-source chat model with parameter-efficient tuning on self-chat data, 2023.
  168. Meta-gradient reinforcement learning. ArXiv, 2018.
  169. Supervised Knowledge Makes Large Language Models Better In-context Learners. arXiv preprint arXiv:2312.15918, 2023.
  170. Tree of thoughts: Deliberate problem solving with large language models. arXiv preprint arXiv:2305.10601, 2023a.
  171. React: Synergizing reasoning and acting in language models. arXiv preprint arXiv:2210.03629, 2023b.
  172. Compositional exemplars for in-context learning. arXiv preprint arXiv:2302.05698, 2023.
  173. Towards accurate model selection in deep unsupervised domain adaptation. In K. Chaudhuri and R. Salakhutdinov (eds.), Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pp.  7124–7133. PMLR, Jun 9–15 2019.
  174. YouWrite. The AI Search Engine You Control. Technical report, You.com, 2022. URL https://you.com.
  175. A survey on neural-symbolic learning systems. Neural Networks, 166:105–126, 2023. ISSN 0893-6080.
  176. Self-rewarding language models. arXiv preprint arXiv:2401.10020, 2024.
  177. Neural, symbolic and neural-symbolic reasoning on knowledge graphs. AIOpen, pp.  14–35, 2021.
Citations (6)

Summary

  • The paper introduces a NeSy framework that combines LLMs' zero- and few-shot learning with solver integration to dynamically construct and evaluate computational graphs.
  • It proposes a novel VERTEX score, inspired by the Fréchet distance, to benchmark multi-step generative processes and complex reasoning tasks.
  • The framework democratizes AI by fusing rule-based symbolic methods with sub-symbolic learning, paving the way for adaptive and autonomous systems.

Introduction

The field of AI has traditionally been divided into two approaches: symbolic AI, which focuses on rule-based systems, and sub-symbolic AI, reliant on neural networks and statistical methods. Bridging these is the domain of Neuro-Symbolic (NeSy) AI, which seeks to combine the strengths of both approaches to unlock new capabilities. The presented work on SymbolicAI introduces a NeSy framework that leverages LLMs within the generative AI landscape. This framework distinguishes itself by incorporating solvers for various domains—ranging from formal language engines to theorem provers—into the generative process, thereby extending the operational range of LLMs and enhancing their utility in complex workflows.

Framework Design

At the core of SymbolicAI is a logic-based philosophy for managing data streams, enabling the dynamic construction and evaluation of computational graphs. SymbolicAI capitalizes on the generative models' zero- and few-shot learning capabilities to model these graphs and uses a varied assembly of solvers for specialized problem-solving. To facilitate this, SymbolicAI defines symbols and expressions capable of morphing into complex structures via polymorphic operations. These operations, along with instruction-based models for fine-tuning LLMs, overarching benchmarking, and a novel quality measure – the VERTEX score, solidify the framework’s utility and robustness. The integration emphasizes the utilization of language as a vehicle for computation, offering a structured path for mapping real-world issues through language inference techniques.

Evaluation Protocol

The proposed VERTEX score tackles the challenge of evaluating multi-step generative processes. This measure is essential for assessing computational graphs and is derived using principles similar to the Fréchet distance in generative models. It captures the trajectory of distributions associated with generative nodes and evaluates them concerning a reference distribution. SymbolicAI claims to facilitate advanced learning protocols and the development of self-referential agents by creating an apparatus for LLMs to manage task execution autonomously. In an essential departure from conventional methods, the framework proposes a computational process devoid of rich-context limitations, enabling long-horizon planning and complex reasoning for machine-based learning models.

Contributions and Implications

SymbolicAI contributes notably to the field by providing a tangible framework that seamlessly integrates solvers with generative models. It operationalizes the construction of computational graphs capable of executing intricate tasks by leveraging LLMs as semantic parsers. Moreover, the framework advocates open-source AI development, which aligns with the broader push for democratizing AI technology.

Conclusion

SymbolicAI has the potential to serve as a foundation for versatile applications through its innovative melding of symbolic and sub-symbolic AI elements. Its introduction of a quality measure, combined with a significant benchmark, poses a substantial step forward in systematizing the evaluation of NeSy processes. Such a framework paves the road to more adaptive, efficient, and autonomous AI systems by enhancing the ability to learn and reason across various contexts and domains.

Github Logo Streamline Icon: https://streamlinehq.com
Youtube Logo Streamline Icon: https://streamlinehq.com
Reddit Logo Streamline Icon: https://streamlinehq.com

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube