Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
144 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Large-scale Graph Representation Learning of Dynamic Brain Connectome with Transformers (2312.14939v1)

Published 4 Dec 2023 in q-bio.NC, cs.CV, cs.LG, and eess.IV

Abstract: Graph Transformers have recently been successful in various graph representation learning tasks, providing a number of advantages over message-passing Graph Neural Networks. Utilizing Graph Transformers for learning the representation of the brain functional connectivity network is also gaining interest. However, studies to date have underlooked the temporal dynamics of functional connectivity, which fluctuates over time. Here, we propose a method for learning the representation of dynamic functional connectivity with Graph Transformers. Specifically, we define the connectome embedding, which holds the position, structure, and time information of the functional connectivity graph, and use Transformers to learn its representation across time. We perform experiments with over 50,000 resting-state fMRI samples obtained from three datasets, which is the largest number of fMRI data used in studies by far. The experimental results show that our proposed method outperforms other competitive baselines in gender classification and age regression tasks based on the functional connectivity extracted from the fMRI data.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (26)
  1. A. Behrouz and M. Seltzer. Anomaly detection in multiplex dynamic networks: from blockchain security to brain disease prediction. arXiv preprint arXiv:2211.08378, 2022.
  2. A. Behrouz and M. Seltzer. Admire++: Explainable anomaly detection in the human brain via inductive learning on temporal multiplex networks. In ICML 3rd Workshop on Interpretable Machine Learning in Healthcare (IMLH), 2023.
  3. Graph neural networks in network neuroscience. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(5):5833–5848, 2022.
  4. R. Botvinik-Nezer and T. D. Wager. Reproducibility in neuroimaging analysis: Challenges and solutions. Biological Psychiatry: Cognitive Neuroscience and Neuroimaging, 2022.
  5. Dyndepnet: Learning time-varying dependency structures from fmri data via dynamic graph structure learning. In ICML 3rd Workshop on Interpretable Machine Learning in Healthcare (IMLH), 2023.
  6. The adolescent brain cognitive development (abcd) study: imaging acquisition across 21 sites. Developmental cognitive neuroscience, 32:43–54, 2018.
  7. Beyond the snapshot: Brain tokenized graph transformer for longitudinal brain functional connectome embedding. arXiv preprint arXiv:2307.00858, 2023.
  8. V. P. Dwivedi and X. Bresson. A generalization of transformer networks to graphs. arXiv preprint arXiv:2012.09699, 2020.
  9. The minimal preprocessing pipelines for the human connectome project. Neuroimage, 80:105–124, 2013.
  10. Functional connectome–based predictive modeling in autism. Biological psychiatry, 92(8):626–642, 2022.
  11. Functional magnetic resonance imaging, volume 1. Sinauer Associates Sunderland, MA, 2004.
  12. Brain network transformer. Advances in Neural Information Processing Systems, 35:25586–25599, 2022.
  13. Understanding graph isomorphism network for rs-fmri functional connectivity analysis. Frontiers in neuroscience, 14:630, 2020.
  14. Learning dynamic graph representation of brain connectome with spatio-temporal attention. Advances in Neural Information Processing Systems, 34:4314–4327, 2021.
  15. Rethinking graph transformers with spectral attention. Advances in Neural Information Processing Systems, 34:21618–21629, 2021.
  16. The uk biobank imaging enhancement of 100,000 participants: rationale, data collection, management and future directions. Nature communications, 11(1):2624, 2020.
  17. Reproducible brain-wide association studies require thousands of individuals. Nature, 603(7902):654–660, 2022.
  18. Transformer for graphs: An overview from architecture perspective. arXiv preprint arXiv:2202.08455, 2022.
  19. On predictability of individual functional connectivity networks from clinical characteristics. Human Brain Mapping, 43(17):5250–5265, 2022.
  20. Attending to graph transformers. arXiv preprint arXiv:2302.04181, 2023.
  21. A survey on oversmoothing in graph neural networks. arXiv preprint arxiv:2303.10993, 2023.
  22. Local-global parcellation of the human cerebral cortex from intrinsic functional connectivity mri. Cerebral Cortex, 28(9):3095–3114, 2017.
  23. L. N. Smith and N. Topin. Super-convergence: Very fast training of neural networks using large learning rates. In Artificial Intelligence and Machine Learning for Multi-Domain Operations Applications, volume 11006, page 1100612. International Society for Optics and Photonics, 2019.
  24. Neuroevolve: A dynamic brain graph deep generative model. In ICML 3rd Workshop on Interpretable Machine Learning in Healthcare (IMLH), 2023.
  25. Attention is all you need. In Advances in neural information processing systems, pages 5998–6008, 2017.
  26. Do transformers really perform badly for graph representation? Advances in Neural Information Processing Systems, 34:28877–28888, 2021.
Citations (1)

Summary

  • The paper introduces TeNeT, a novel graph transformer method that captures dynamic brain connectivity by integrating structural, positional, and time embeddings from fMRI data.
  • It leverages a two-step transformer architecture with spatial and temporal modules to effectively process dynamic connectome graphs, outperforming traditional GNN and static transformer baselines.
  • Experiments on over 50,000 samples demonstrate superior performance in gender classification (AUROC) and age regression (R²), underscoring its scalability and robustness.

The paper "Large-scale Graph Representation Learning of Dynamic Brain Connectome with Transformers" (2312.14939) introduces TeNeT, a novel Graph Transformer (GT)-based method designed to learn representations of dynamic functional connectivity (FC) from resting-state fMRI data. The core problem addressed is the limitation of existing methods, such as Graph Neural Networks (GNNs) and static GTs, in effectively capturing the temporal fluctuations of brain FC and demonstrating performance on large-scale datasets, which is crucial for generalizability.

TeNeT tackles this by processing fMRI data as a sequence of dynamic FC graphs across time. The practical implementation involves several key steps:

  1. Data Preprocessing: Start with the ROI-timeseries matrix extracted from 4D fMRI data. The paper uses the Schaefer atlas with 400 ROIs.
  2. Connectome Embedding: For each time point tt, a connectome embedding Et\mathcal{E}_t is constructed. This embedding combines three types of information:
    • Structure Embedding: Calculated from the dynamic FC graph at time tt. This dynamic FC is typically derived using a sliding-window approach on the ROI-timeseries, computing correlation coefficients within a window of length Γ\Gamma and stride SS. The correlation matrix Cˉt\bar{\mathcal{C}}_t represents edge weights.
    • Position Embedding: Derived from the structure embedding by combining the correlation matrix (after removing self-loops) with an identity matrix. This combined graph embedding is then projected to a lower dimension DD using a two-layer MLP.
    • Time Embedding: Obtained from the ROI-timeseries within the sliding window using a GRU, capturing temporal context.
    • The final connectome embedding Et\mathcal{E}_t is formed by concatenating the MLP-processed graph embedding and the GRU-derived time embedding for each time point. This results in a sequence of embeddings (E1,E2,...,ET)(\mathcal{E}_1, \mathcal{E}_2, ..., \mathcal{E}_T).
  3. TeNeT Architecture: The model uses a two-step Transformer composition f=hgf = h \circ g.
    • Spatial Transformer (gg): A Connectome Transformer processes each time point's connectome embedding Et\mathcal{E}_t independently using self-attention. This layer is augmented by injecting structural information from the dynamic graph, specifically the 1-hop connectivity matrix Cˉt\bar{\mathcal{C}}_t and node degree information. A learnable token is prepended to the embedding sequence at each time point. After LL layers, the token vector for each time point tt (vt\mathbf{v}_t) captures spatially-attended features at that specific moment.
    • Temporal Transformer (hh): A standard Transformer Encoder takes the sequence of token vectors (v1,v2,...,vT)(\mathbf{v}_1, \mathbf{v}_2, ..., \mathbf{v}_T) generated by the spatial transformer as input. It applies self-attention across the time dimension to learn the dynamic patterns. Again, a learnable token is used. After LL layers, the final token vector Edyn\mathcal{E}_{\text{dyn}} represents the comprehensive dynamic FC representation for the entire fMRI sample.
  4. Downstream Tasks: The final token vector Edyn\mathcal{E}_{\text{dyn}} is used as input to a simple classifier (for gender) or regressor (for age) head.

The authors conducted extensive experiments using over 50,000 resting-state fMRI samples from three large datasets (UKB, ABCD, HCP subsets), demonstrating the method's capability on an unprecedented scale for fMRI studies. Performance was evaluated on gender classification (AUROC) and age regression (R2R^2).

Implementation Details & Considerations:

  • Dataset Size: The paper highlights the use of large datasets (>50,000 samples) as crucial for enhancing replicability and generalizability, addressing a major challenge in neuroimaging. Implementing this requires significant data storage and processing capabilities.
  • Hyperparameters: The model configuration included 4 layers with a hidden dimension of 1024. Training utilized the Adam optimizer with a one-cycle learning rate schedule. Hyperparameter tuning (batch size, learning rate) was performed using grid search.
  • Computational Requirements: Training on large datasets requires substantial computational resources. The experiments were run on an NVIDIA GeForce RTX 3090, indicating the need for high-end GPUs, potentially multiple for larger-scale training or inference.
  • Dynamic FC Calculation: The sliding-window approach introduces choices for window length (Γ\Gamma) and stride (SS). These parameters impact the temporal resolution and the number of dynamic graphs generated per subject.
  • Memory Usage: Processing sequences of graph embeddings can be memory-intensive, especially with large numbers of ROIs (nodes) and time points. Batch sizing needs to be carefully considered based on available GPU memory.
  • Ablation Studies: The ablation results emphasize the importance of both the GRU-derived time encoding and the dynamic nature of the graph features, suggesting that removing either component degrades performance. This guides implementation choices regarding input features.
  • Scalability: While demonstrated on large datasets, scaling to even larger cohorts or higher resolution atlases would require further optimization of memory usage and distributed training strategies.

Practical Applications:

TeNeT can be applied to extract powerful representations of brain connectivity dynamics for various tasks:

  • Predicting Phenotypes: As demonstrated, predicting demographic traits like gender and age from resting-state fMRI data.
  • Clinical Diagnosis/Prognosis: Applying the model to predict clinical phenotypes, such as neurological or psychiatric disorders, which are often associated with altered dynamic FC patterns.
  • Biomarker Discovery: The learned dynamic representations could potentially serve as biomarkers for disease states or treatment response.
  • Understanding Brain Dynamics: Future work on interpreting the attention weights within TeNeT could provide insights into which brain regions and temporal patterns are most informative for specific tasks or conditions.

The comparative results show TeNeT's superiority over static and dynamic GNN baselines, and a static GT baseline, supporting the value of its approach to capturing dynamic FC with Transformers on large datasets. The model's design, integrating position, structure, and time information within a Transformer framework, offers a promising direction for leveraging complex neuroimaging data.