Discrete, compositional, and symbolic representations through attractor dynamics (2310.01807v2)
Abstract: Symbolic systems are powerful frameworks for modeling cognitive processes as they encapsulate the rules and relationships fundamental to many aspects of human reasoning and behavior. Central to these models are systematicity, compositionality, and productivity, making them invaluable in both cognitive science and artificial intelligence. However, certain limitations remain. For instance, the integration of structured symbolic processes and latent sub-symbolic processes has been implemented at the computational level through fiat methods such as quantization or softmax sampling, which assume, rather than derive, the operations underpinning discretization and symbolicization. In this work, we introduce a novel neural stochastic dynamical systems model that integrates attractor dynamics with symbolic representations to model cognitive processes akin to the probabilistic language of thought (PLoT). Our model segments the continuous representational space into discrete basins, with attractor states corresponding to symbolic sequences, that reflect the semanticity and compositionality characteristic of symbolic systems through unsupervised learning, rather than relying on pre-defined primitives. Moreover, like PLoT, our model learns to sample a diverse distribution of attractor states that reflect the mutual information between the input data and the symbolic encodings. This approach establishes a unified framework that integrates both symbolic and sub-symbolic processing through neural dynamics, a neuro-plausible substrate with proven expressivity in AI, offering a more comprehensive model that mirrors the complex duality of cognitive operations.
- Jerry A Fodor. The language of thought. Crowell Press, 1975.
- Concepts in a probabilistic language of thought. Technical report, Center for Brains, Minds and Machines (CBMM), 2014.
- Building machines that learn and think like people. Behavioral and Brain Sciences, 40, 2017.
- Yoshua Bengio. The consciousness prior. arXiv preprint arXiv:1709.08568, 2019.
- Inductive biases for deep learning of higher-level cognition. Proceedings of the Royal Society A, 478(2266):20210068, 2022.
- Symbols and mental programs: a hypothesis about human singularity. Trends in Cognitive Sciences, 2022.
- Connectionism and cognitive architecture: A critical analysis. Cognition, 28(1-2):3–71, 1988.
- David Marr. Vision: A computational investigation into the human representation and processing of visual information, 1982.
- James L McClelland. Emergence in cognitive science. Topics in cognitive science, 2(4):751–770, 2010.
- Neural discrete representation learning. Neural Information Processing Systems (NIPS), 2017.
- Simplicial embeddings in self-supervised learning and downstream classification. International Conference on Learning Representations (ICLR), 2023.
- From word models to world models: Translating from natural language to the probabilistic language of thought. arXiv preprint arXiv:2306.12672, 2023.
- Learning compositional neural programs for continuous control. arXiv preprint arXiv:2007.13363, 2021.
- Neural systematic binder. International Conference on Learning Representations, 2023.
- DreamCoder: Bootstrapping inductive program synthesis with wake-sleep library learning. Programming Language Design and Implementation, 2021.
- Neural production systems: Learning rule-governed visual dynamics. Neural Information Processing Systems (NeurIPS), 2021.
- Recurrent independent mechanisms. International Conference on Learning Representations (ICLR), 2021.
- Object files and schemata: Factorizing declarative and procedural knowledge in dynamical systems. International Conference on Learning Representations (ICLR), 2021.
- Coordination among neural modules through a shared global workspace. International Conference on Learning Representations (ICLR), 2022.
- Is a modular architecture enough? Neural Information Processing Systems (NeurIPS), 2022.
- Sources of richness and ineffability for phenomenally conscious states. arXiv preprint arXiv:2302.06403, 2023.
- Gflownet foundations. Journal of Machine Learning Research, 24(210):1–55, 2023.
- GFlowNet-EM for learning compositional latent variable models. International Conference on Machine Learning (ICML), 2023.
- Neural stochastic differential equations: Deep latent Gaussian models in the diffusion limit. arXiv preprint arXiv:1905.09883, 2019.
- A theory of continuous generative flow networks. International Conference on Machine Learning (ICML), 2023.
- dsprites: Disentanglement testing sprites dataset. https://github.com/deepmind/dsprites-dataset/, 2017.
- Learning structured output representation using deep conditional generative models. Neural Information Processing Systems (NIPS), 2015.
- Better training of GFlowNets with local credit and incomplete trajectories. International Conference on Machine Learning (ICML), 2023.
- Auto-encoding variational Bayes. International Conference on Learning Representations (ICLR), 2014.