Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Multimodal foundation models are better simulators of the human brain (2208.08263v1)

Published 17 Aug 2022 in cs.NE, cs.AI, and cs.MM

Abstract: Multimodal learning, especially large-scale multimodal pre-training, has developed rapidly over the past few years and led to the greatest advances in AI. Despite its effectiveness, understanding the underlying mechanism of multimodal pre-training models still remains a grand challenge. Revealing the explainability of such models is likely to enable breakthroughs of novel learning paradigms in the AI field. To this end, given the multimodal nature of the human brain, we propose to explore the explainability of multimodal learning models with the aid of non-invasive brain imaging technologies such as functional magnetic resonance imaging (fMRI). Concretely, we first present a newly-designed multimodal foundation model pre-trained on 15 million image-text pairs, which has shown strong multimodal understanding and generalization abilities in a variety of cognitive downstream tasks. Further, from the perspective of neural encoding (based on our foundation model), we find that both visual and lingual encoders trained multimodally are more brain-like compared with unimodal ones. Particularly, we identify a number of brain regions where multimodally-trained encoders demonstrate better neural encoding performance. This is consistent with the findings in existing studies on exploring brain multi-sensory integration. Therefore, we believe that multimodal foundation models are more suitable tools for neuroscientists to study the multimodal signal processing mechanisms in the human brain. Our findings also demonstrate the potential of multimodal foundation models as ideal computational simulators to promote both AI-for-brain and brain-for-AI research.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. Haoyu Lu (24 papers)
  2. Qiongyi Zhou (4 papers)
  3. Nanyi Fei (14 papers)
  4. Zhiwu Lu (51 papers)
  5. Mingyu Ding (82 papers)
  6. Jingyuan Wen (5 papers)
  7. Changde Du (25 papers)
  8. Xin Zhao (160 papers)
  9. Hao Sun (383 papers)
  10. Huiguang He (26 papers)
  11. Ji-Rong Wen (299 papers)
Citations (12)