Multi-Modal Cognitive Maps based on Neural Networks trained on Successor Representations (2401.01364v1)
Abstract: Cognitive maps are a proposed concept on how the brain efficiently organizes memories and retrieves context out of them. The entorhinal-hippocampal complex is heavily involved in episodic and relational memory processing, as well as spatial navigation and is thought to built cognitive maps via place and grid cells. To make use of the promising properties of cognitive maps, we set up a multi-modal neural network using successor representations which is able to model place cell dynamics and cognitive map representations. Here, we use multi-modal inputs consisting of images and word embeddings. The network learns the similarities between novel inputs and the training database and therefore the representation of the cognitive map successfully. Subsequently, the prediction of the network can be used to infer from one modality to another with over $90\%$ accuracy. The proposed method could therefore be a building block to improve current AI systems for better understanding of the environment and the different modalities in which objects appear. The association of specific modalities with certain encounters can therefore lead to context awareness in novel situations when similar encounters with less information occur and additional information can be inferred from the learned cognitive map. Cognitive maps, as represented by the entorhinal-hippocampal complex in the brain, organize and retrieve context from memories, suggesting that LLMs like ChatGPT could harness similar architectures to function as a high-level processing center, akin to how the hippocampus operates within the cortex hierarchy. Finally, by utilizing multi-modal inputs, LLMs can potentially bridge the gap between different forms of data (like images and words), paving the way for context-awareness and grounding of abstract concepts through learned associations, addressing the grounding problem in AI.
- Dissociating language and thought in large language models: a cognitive perspective. arxiv, 2023.
- Gary Marcus. Deep learning: A critical appraisal. arXiv preprint arXiv:1801.00631, 2018.
- Bertram Opitz. Memory function and the hippocampus. Frontiers of Neurology and Neuroscience, 34:51–59, 2014.
- The cognitive map in humans: spatial navigation and beyond. Nature neuroscience, 20(11):1504–1513, 2017.
- Inferences on a multidimensional social hierarchy use a grid-like code. bioRxiv, pages 2020–05, 2021.
- Grid cells map the visual world. Nature Neuroscience, 21(2), 2018.
- Edmund T Rolls. Limbic systems for emotion and for memory, but no single limbic system. cortex, 62:119–157, 2015.
- The hippocampus as a spatial map: preliminary evidence from unit activity in the freely-moving rat. Brain research, 1971.
- Microstructure of a spatial map in the entorhinal cortex. Nature, 436(7052):801–806, 2005.
- Fragmentation of grid cell maps in a multicompartment environment. Nature Neuroscience, 12(10):1325–1332, October 2009.
- The hippocampus as a cognitive map. Oxford university press, 1978.
- Spatial representation in the hippocampal formation: a history. Nature neuroscience, 20(11):1448–1464, 2017.
- Navigating cognition: Spatial codes for human thinking. Science, 362(6415), 2018.
- Memory hierarchies map onto the hippocampal long axis in humans. Nature neuroscience, 18(11):1562–1564, 2015.
- The tolman-eichenbaum machine: Unifying space and relational memory through generalization in the hippocampal formation. Cell, 183(5):1249–1263, 2020.
- Design principles of the hippocampal cognitive map. Advances in neural information processing systems, 27:2528–2536, 2014.
- The hippocampus as a predictive map. Nature neuroscience, 20(11):1643, 2017.
- Neural network based successor representations to form cognitive maps of space and language. Scientific Reports, 12:11233, 2022. Number: 1 Publisher: Nature Publishing Group.
- Neural network based formation of cognitive maps of semantic spaces and the putative emergence of abstract concepts. Scientific Reports, 13(1):3644, 2023.
- Word class representations spontaneously emerge in a deep neural network trained on next word prediction. arXiv preprint arXiv:2302.07588, 2023.
- Conceptual cognitive maps formation with neural successor networks and word embeddings, 2023.
- Peter Dayan. Improving Generalization for Temporal Difference Learning: The Successor Representation. Neural Computation, 5(4):613–624, 07 1993.
- Li Deng. The mnist database of handwritten digit images for machine learning research. IEEE Signal Processing Magazine, 29(6):141–142, 2012.
- spaCy 2: Natural language understanding with Bloom embeddings, convolutional neural networks and incremental parsing. To appear, 2017.
- Laurens Van der Maaten and Geoffrey Hinton. Visualizing data using t-sne. Journal of machine learning research, 9(11), 2008.
- How to use t-sne effectively. Distill, 1(10):e2, 2016.
- Catalina A Vallejos. Exploring a world of a thousand dimensions. Nature biotechnology, 37(12):1423–1424, 2019.
- Visualizing structure and transitions in high-dimensional biological data. Nature biotechnology, 37(12):1482–1492, 2019.
- Warren S Torgerson. Multidimensional scaling: I. theory and method. Psychometrika, 17(4):401–419, 1952.
- Joseph B Kruskal. Nonmetric multidimensional scaling: a numerical method. Psychometrika, 29(2):115–129, 1964.
- Joseph B Kruskal. Multidimensional scaling. Number 11. Sage, 1978.
- Multidimensional scaling. In Handbook of data visualization, pages 315–347. Springer, 2008.
- Analysis of continuous neuronal activity evoked by natural speech with computational corpus linguistics methods. Language, Cognition and Neuroscience, 36(2):167–186, 2021.
- Quantifying the separability of data classes in neural networks. Neural Networks, 139:278–293, 2021.
- Analysis and visualization of sleep stages based on deep neural networks. Neurobiology of sleep and circadian rhythms, 10:100064, 2021.
- Analysis of structure and dynamics in three-neuron motifs. Frontiers in Computational Neuroscience, 13:5, 2019.
- Recurrence resonance” in three-neuron motifs. Frontiers in computational neuroscience, page 64, 2019.
- Weight statistics controls dynamics in recurrent neural networks. PloS one, 14(4):e0214541, 2019.
- Dynamics and information import in recurrent neural networks. Frontiers in Computational Neuroscience, 16, 2022.
- A statistical method for analyzing and comparing spatiotemporal cortical activation patterns. Scientific reports, 8(1):1–9, 2018.
- Analysis of multichannel eeg patterns during human sleep: a novel approach. Frontiers in human neuroscience, 12:121, 2018.
- Microstructure of cortical activity during sleep reflects respiratory events and state of daytime vigilance. Somnologie, 23(2):72–79, 2019.
- François Chollet et al. Keras, 2015.
- TensorFlow: Large-scale machine learning on heterogeneous systems, 2015. Software available from tensorflow.org.
- Array programming with NumPy. Nature, 585(7825):357–362, September 2020.
- Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825–2830, 2011.
- J. D. Hunter. Matplotlib: A 2d graphics environment. Computing in Science & Engineering, 9(3):90–95, 2007.
- Graph transformer networks. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc., 2019.
- Jurgen Schmidhuber. Towards compositional learning in dynamic networkstechnical report. 1990.
- Jurgen Schmidhuber. Learning to generate subgoals for action sequences. IJCNN-91-Seattle International Joint Conference on Neural Networks, ii:453 vol.2–, 1991.
- Planning simple trajectories using neural subgoal generators. In Proceedings of the Second International Conference on From Animals to Animats 2: Simulation of Adaptive Behavior: Simulation of Adaptive Behavior, page 196–202, Cambridge, MA, USA, 1993. MIT Press.
- Juergen Schmidhuber. On learning to think: Algorithmic information theory for novel combinations of reinforcement learning controllers and recurrent neural world models, 2015.
- Juergen Schmidhuber. One big net for everything, 2018.
- Cade Metz. The godfather of ai leaves google and warns of danger ahead. New York Times, 2023.
- Eliezer Yudkowsky. Pausing ai developments isn’t enough. we need to shut it all down. Time Magazine (Mar, 2023). https://time. com/6266923/ai-eliezer-yudkowsky-open-letter-not-enough, 2023.
- Explaining the black-box model: A survey of local interpretation methods for deep neural networks. Neurocomputing, 419:168–182, 2021.
- Michael C. Frank. Baby steps in evaluating the capacities of large language models. Nature Reviews Psychology, June 2023.
- Bias in data-driven artificial intelligence systems—an introductory survey. WIREs Data Mining and Knowledge Discovery, 10(3):e1356, 2020.
- Towards a standard for identifying and managing bias in artificial intelligence. NIST special publication, 1270(10.6028), 2022.
- Luke Munn. The uselessness of AI ethics. AI and Ethics, 3(3):869–877, August 2023.
- Paul Stoewer (5 papers)
- Achim Schilling (34 papers)
- Andreas Maier (394 papers)
- Patrick Krauss (40 papers)