Joint-Embedding Masked Autoencoder for Self-supervised Learning of Dynamic Functional Connectivity from the Human Brain (2403.06432v2)
Abstract: Graph Neural Networks (GNNs) have shown promise in learning dynamic functional connectivity for distinguishing phenotypes from human brain networks. However, obtaining extensive labeled clinical data for training is often resource-intensive, making practical application difficult. Leveraging unlabeled data thus becomes crucial for representation learning in a label-scarce setting. Although generative self-supervised learning techniques, especially masked autoencoders, have shown promising results in representation learning in various domains, their application to dynamic graphs for dynamic functional connectivity remains underexplored, facing challenges in capturing high-level semantic representations. Here, we introduce the Spatio-Temporal Joint Embedding Masked Autoencoder (ST-JEMA), drawing inspiration from the Joint Embedding Predictive Architecture (JEPA) in computer vision. ST-JEMA employs a JEPA-inspired strategy for reconstructing dynamic graphs, which enables the learning of higher-level semantic representations considering temporal perspectives, addressing the challenges in fMRI data representation learning. Utilizing the large-scale UK Biobank dataset for self-supervised learning, ST-JEMA shows exceptional representation learning performance on dynamic functional connectivity demonstrating superiority over previous methods in predicting phenotypes and psychiatric diagnoses across eight benchmark fMRI datasets even with limited samples and effectiveness of temporal reconstruction on missing data scenarios. These findings highlight the potential of our approach as a robust representation learning method for leveraging label-scarce fMRI data.
- G. Alain and Y. Bengio (2016) Understanding intermediate layers using linear classifier probes. arXiv preprint arXiv:1610.01644. Cited by: §VI-B.
- Self-supervised learning from images with a joint-embedding predictive architecture. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 15619–15629. Cited by: §I, §II-C, §III-D, §IV-A, §IV-B, §IV-C, §VII.
- The lifespan human connectome project in aging: an overview. Neuroimage 185, pp. 335–348. Cited by: §V.
- ADHD-200 global competition: diagnosing adhd using personal characteristic data can outperform resting state fmri measurements. Frontiers in systems neuroscience 6, pp. 69. Cited by: §V.
- The adolescent brain cognitive development (abcd) study: imaging acquisition across 21 sites. Developmental cognitive neuroscience 32, pp. 43–54. Cited by: §V.
- An empirical study of training self-supervised vision transformers. in 2021 ieee. In CVF International Conference on Computer Vision (ICCV), pp. 9620–9629. Cited by: §I.
- A generative self-supervised framework using functional connectivity in fmri data. In Temporal Graph Learning Workshop@ NeurIPS 2023, Cited by: §I, §I, §II-B, §IV-D, §IV-D, §V.
- Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555. Cited by: §III-C.
- The neuro bureau preprocessing initiative: open sharing of preprocessed neuroimaging data and derivatives. Frontiers in Neuroinformatics 7 (27), pp. 5. Cited by: §V.
- Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805. Cited by: §I.
- Spatio-temporal graph convolution for resting-state fmri analysis. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2020: 23rd International Conference, Lima, Peru, October 4–8, 2020, Proceedings, Part VII 23, pp. 528–538. Cited by: §I, §II-B.
- Bootstrap your own latent-a new approach to self-supervised learning. Advances in neural information processing systems 33, pp. 21271–21284. Cited by: §IV-A.
- Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 16000–16009. Cited by: §I, §II-C.
- Graphmae: self-supervised masked graph autoencoders. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp. 594–604. Cited by: §I, §I, §I, §II-B, §IV-C, §V.
- Fbnetgen: task-aware gnn-based fmri analysis via functional brain network generation. In International Conference on Medical Imaging with Deep Learning, pp. 618–637. Cited by: §II-A.
- 3D convolutional neural networks for classification of functional connectomes. In International Workshop on Deep Learning in Medical Image Analysis, pp. 137–145. Cited by: §I.
- Learning dynamic graph representation of brain connectome with spatio-temporal attention. Advances in Neural Information Processing Systems 34, pp. 4314–4327. Cited by: §I, §II-A, §III-B, §IV-B, §V.
- B. Kim and J. C. Ye (2020) Understanding graph isomorphism network for rs-fmri functional connectivity analysis. Frontiers in neuroscience 14, pp. 630. Cited by: §I.
- T. N. Kipf and M. Welling (2016) Variational graph auto-encoders. arXiv preprint arXiv:1611.07308. Cited by: §II-B, §IV-C, §V.
- Neuroprogression across the early course of psychosis. Journal of psychiatry and brain science 5. Cited by: §V.
- Braingnn: interpretable brain graph neural network for fmri analysis. Medical Image Analysis 74, pp. 102233. Cited by: §II-A.
- BrainTGL: a dynamic graph representation learning model for brain network analysis. Computers in Biology and Medicine 153, pp. 106521. Cited by: §II-A.
- Functional imaging of the hemodynamic sensory gating response in schizophrenia. Human brain mapping 34 (9), pp. 2302–2312. Cited by: §V.
- Spatio-temporal deep graph infomax. arXiv preprint arXiv:1904.06316. Cited by: §V.
- Local-global parcellation of the human cerebral cortex from intrinsic functional connectivity mri. Cerebral cortex 28 (9), pp. 3095–3114. Cited by: §III-C.
- The lifespan human connectome project in development: a large-scale study of brain connectivity development in 5–21 year olds. Neuroimage 183, pp. 456–468. Cited by: §V.
- UK biobank: an open access resource for identifying the causes of a wide range of complex diseases of middle and old age. PLoS medicine 12 (3), pp. e1001779. Cited by: §V.
- Mgae: masked autoencoders for self-supervised learning on graphs. arXiv preprint arXiv:2201.02534. Cited by: §IV-C.
- Mlp-mixer: an all-mlp architecture for vision. Advances in neural information processing systems 34, pp. 24261–24272. Cited by: §IV-C.
- Deep graph infomax. arXiv preprint arXiv:1809.10341. Cited by: §II-B, §V.
- Modeling dynamic characteristics of brain functional connectivity networks using resting-state functional mri. Medical image analysis 71, pp. 102063. Cited by: §II-A.
- Masked feature prediction for self-supervised visual pre-training. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 14668–14678. Cited by: §I.
- H. WU-Minn (2017) 1200 subjects data release reference manual. URL https://www. humanconnectome. org 565. Cited by: §V.
- Simgrace: a simple framework for graph contrastive learning without data augmentation. In Proceedings of the ACM Web Conference 2022, pp. 1070–1079. Cited by: §I, §II-B, §V.
- How powerful are graph neural networks?. International Conference on Learning Representations. Cited by: §V.
- Groupinn: grouping-based interpretable neural network for classification of limited, noisy brain data. In Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining, pp. 772–782. Cited by: §II-A.
- Functional connectivity magnetic resonance imaging classification of autism spectrum disorder using the multisite abide dataset. In 2019 IEEE EMBS International Conference on Biomedical & Health Informatics (BHI), pp. 1–4. Cited by: §I.
- Graph contrastive learning with augmentations. Advances in neural information processing systems 33, pp. 5812–5823. Cited by: §I, §II-B.
- Are graph augmentations necessary? simple graph contrastive learning for recommendation. In Proceedings of the 45th international ACM SIGIR conference on research and development in information retrieval, pp. 1294–1303. Cited by: §II-B.
- Layer normalization. arXiv preprint arXiv:1607.06450. Cited by: §D-A.
- The lifespan human connectome project in aging: an overview. Neuroimage 185, pp. 335–348. Cited by: §B-B.
- ADHD-200 global competition: diagnosing adhd using personal characteristic data can outperform resting state fmri measurements. Frontiers in systems neuroscience 6, pp. 69. Cited by: §B-B.
- The adolescent brain cognitive development (abcd) study: imaging acquisition across 21 sites. Developmental cognitive neuroscience 32, pp. 43–54. Cited by: §B-B.
- A generative self-supervised framework using functional connectivity in fmri data. In Temporal Graph Learning Workshop@ NeurIPS 2023, Cited by: Appendix C.
- The neuro bureau preprocessing initiative: open sharing of preprocessed neuroimaging data and derivatives. Frontiers in Neuroinformatics 7 (27), pp. 5. Cited by: §B-B.
- Understanding and improving the role of projection head in self-supervised learning. arXiv preprint arXiv:2212.11491. Cited by: §F-A.
- D. Hendrycks and K. Gimpel (2016) Gaussian error linear units (gelus). arXiv preprint arXiv:1606.08415. Cited by: §D-A.
- Graphmae: self-supervised masked graph autoencoders. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp. 594–604. Cited by: Appendix C, §F-B.
- Squeeze-and-excitation networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7132–7141. Cited by: §D-A.
- Understanding dimensional collapse in contrastive self-supervised learning. International Conference on Learning Representations. Cited by: §F-A.
- Learning dynamic graph representation of brain connectome with spatio-temporal attention. Advances in Neural Information Processing Systems 34, pp. 4314–4327. Cited by: §B-B, §D-A, §D-A, §D-A, §D-B.
- B. Kim and J. C. Ye (2020) Understanding graph isomorphism network for rs-fmri functional connectivity analysis. Frontiers in neuroscience 14, pp. 630. Cited by: §D-A.
- D. P. Kingma and M. Welling (2013) Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114. Cited by: §F-B.
- T. N. Kipf and M. Welling (2016) Variational graph auto-encoders. arXiv preprint arXiv:1611.07308. Cited by: Appendix C, Appendix C, §F-B.
- A. Leman and B. Weisfeiler (1968) A reduction of a graph to a canonical form and an algebra arising during this reduction. Nauchno-Technicheskaya Informatsiya 2 (9), pp. 12–16. Cited by: §D-A.
- Neuroprogression across the early course of psychosis. Journal of psychiatry and brain science 5. Cited by: §B-B.
- Functional imaging of the hemodynamic sensory gating response in schizophrenia. Human brain mapping 34 (9), pp. 2302–2312. Cited by: §B-B.
- Spatio-temporal deep graph infomax. arXiv preprint arXiv:1904.06316. Cited by: Appendix C.
- Deepwalk: online learning of social representations. In Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining, pp. 701–710. Cited by: §F-B.
- Local-global parcellation of the human cerebral cortex from intrinsic functional connectivity mri. Cerebral cortex 28 (9), pp. 3095–3114. Cited by: §B-A.
- The lifespan human connectome project in development: a large-scale study of brain connectivity development in 5–21 year olds. Neuroimage 183, pp. 456–468. Cited by: §B-B.
- UK biobank: an open access resource for identifying the causes of a wide range of complex diseases of middle and old age. PLoS medicine 12 (3), pp. e1001779. Cited by: §B-B.
- Mlp-mixer: an all-mlp architecture for vision. Advances in neural information processing systems 34, pp. 24261–24272. Cited by: §D-A, §D-A, §E-C, TABLE VIII.
- Deep graph infomax. arXiv preprint arXiv:1809.10341. Cited by: Appendix C, §F-A.
- H. WU-Minn (2017) 1200 subjects data release reference manual. URL https://www. humanconnectome. org 565. Cited by: §B-B.
- Simgrace: a simple framework for graph contrastive learning without data augmentation. In Proceedings of the ACM Web Conference 2022, pp. 1070–1079. Cited by: Appendix C, §F-A.
- How powerful are graph neural networks?. International Conference on Learning Representations. Cited by: §D-A, §D-A, TABLE VIII.
- Graph contrastive learning with augmentations. Advances in neural information processing systems 33, pp. 5812–5823. Cited by: §F-A.
- Are graph augmentations necessary? simple graph contrastive learning for recommendation. In Proceedings of the 45th international ACM SIGIR conference on research and development in information retrieval, pp. 1294–1303. Cited by: §F-A.