- The paper introduces Brain-like Space and quantitatively maps AI models' attention patterns to human brain networks, revealing an arc-shaped gradient of brain-likeness.
- The study employs graph theoretic metrics and PCA clustering to compare 151 Transformer models, showing that pretraining paradigms and positional encodings significantly impact brain-likeness.
- The work highlights that brain-likeness, while reflecting high-level network alignment, is decoupled from downstream task performance, guiding future model design and evaluation.
A Unified Geometric Space Bridging AI Models and the Human Brain
Introduction
This paper introduces the concept of "Brain-like Space," a unified geometric framework for situating and comparing the intrinsic organizational topology of AI models with canonical human functional brain networks. The approach leverages graph-theoretic similarity measures to map the spatial attention patterns of Transformer-based models onto seven functional brain networks derived from resting-state fMRI. The study encompasses 151 Transformer-based models, including large vision models (LVMs), LLMs, and large multimodal models (LMMs), and reveals a continuous arc-shaped geometry within this space, reflecting a gradient of brain-likeness. The analysis demonstrates that brain-likeness is not solely determined by model modality but is systematically influenced by pretraining paradigms and positional encoding schemes.
Construction of Brain-like Space
The methodology involves several key steps:
- Functional Brain Network Extraction: A group-level functional connectivity matrix is computed from rs-fMRI data of 1042 subjects, parcellated into seven canonical networks: limbic (LIM), visual (VIS), somatomotor (SMN), dorsal attention (DAN), ventral attention (VAN), frontoparietal (FPN), and default mode (DMN).
- Spatial Attention Graphs in AI Models: For each attention head in a Transformer model, a spatial attention graph is constructed, with nodes representing spatial patches and edge weights derived from attention scores.
- Graph-Theoretic Metrics: Five metrics—average clustering coefficient, modularity, degree standard deviation, average shortest path length, and global efficiency—are extracted for both brain and model graphs.
- Similarity Computation: Cosine similarity between the feature vectors of model attention heads and brain networks yields a seven-dimensional representation for each attention head.
- Dimensionality Reduction and Clustering: PCA projects the seven-dimensional space to two dimensions, explaining 96.07% of the variance. K-means clustering (k=4) segments attention heads into clusters reflecting increasing brain-likeness.
Key Findings
Structured Distribution and Model Categories
- Language-dominant models (LLM, LLM-RoPE, LMM-language, LMM-language-RoPE) are highly concentrated in the most brain-like cluster (C4), with up to 89.2% of attention heads assigned to C4.
- Vision-dominant models show greater heterogeneity; those emphasizing global semantic abstraction (e.g., ViT-Variants-global-semantic) are more brain-like, while local reconstructive models (e.g., DeiT3, MAE) are less so.
- Multimodal models exhibit distinct patterns; RoPE-based positional encoding in LMMs facilitates deep fusion and increases brain-likeness in both language and vision components.
Influence of Pretraining Paradigms
- Data Augmentation: Strategies like AugReg (Mixup, RandAugment) promote global disruption, driving models toward invariance to local distortions and enhancing matches with higher-order cognitive networks. In contrast, 3-Augment (DeiT3) focuses on local stability, limiting brain-likeness.
- Training Objectives: Semantic abstraction objectives (DINO, DINOv3, BEiT, BEiTv2) yield high brain-likeness and strong matches with FPN and DMN. Detail reconstruction objectives (MAE, DINOv2) bias models toward VIS, reducing brain-likeness.
- Distillation: CNN-based teacher distillation (DeiT) suppresses global attention mechanisms, shifting models toward local inductive bias and reducing brain-likeness, especially as model scale increases.
Positional Encoding Schemes
- RoPE-based LLMs and LMMs: RoPE enables a unified geometric prior, facilitating deep cross-modal fusion and increasing brain-likeness, particularly in vision components of multimodal models.
- Learnable Positional Encoding: Models like CLIP and BLIP exhibit functional localization, with vision components less brain-like and language components highly brain-like, reflecting a division of labor.
- A positive but non-significant correlation (Pearson's r = 0.266, p = 0.1555) exists between brain-likeness scores and ImageNet-1k Top-1 accuracy across 30 vision models. Models optimized for engineering efficiency or robustness may diverge from brain-like organization.
Biological Plausibility and Hierarchical Trends
Layer-wise analysis reveals a hierarchical matching pattern: shallow layers align with LIM and VIS, while deeper layers match DAN, VAN, and DMN, mirroring the principal gradient of cortical organization. Larger models develop partial analogs of high-level cognitive processes, but only when scaling is compatible with the training objective.
Implications for Model Design and Evaluation
- Pretraining Paradigm as Meta-Regulator: Global and semantic-level modeling in pretraining enhances brain-likeness and organizational similarity with higher-order networks. Local detail-focused objectives suppress brain-like structures.
- Model Scale: Scaling acts as a catalyst for brain-likeness only when paired with compatible pretraining; otherwise, increased parameters may dilute brain-like efficiency.
- Brain-likeness as an Independent Metric: Brain-likeness should be incorporated as an organizational-interpretive metric, independent of downstream task performance, to enhance model explainability and guide architecture design.
Graph-based Approach and Generalizability
The graph-based framework enables direct, modality-agnostic comparison of AI models and brain networks, facilitating cost-effective and generalizable brain-likeness assessment. The seven-dimensional Brain-like Space and derived brain-likeness score provide operational tools for model evaluation and architectural optimization.
Limitations and Future Directions
- The current approach focuses on spatial attention; incorporating feature channel interactions could yield a more comprehensive assessment.
- Finer-grained brain network atlases may improve correspondence.
- Extension to non-Transformer architectures (MLPs, CNNs) is needed.
- Dynamic analysis of brain-likeness evolution during reasoning remains an open direction.
Conclusion
This study establishes Brain-like Space as a unified geometric framework for quantifying and comparing the intrinsic organization of AI models and the human brain. The findings demonstrate that brain-likeness emerges from the interplay of architecture, pretraining paradigm, and positional encoding, and is not inherently coupled with downstream task performance. The proposed framework offers a principled approach for advancing a unified science of intelligence, bridging the gap between artificial and biological systems.