FIMP: Foundation Model-Informed Message Passing for Graph Neural Networks (2210.09475v5)
Abstract: Foundation models have achieved remarkable success across many domains, relying on pretraining over vast amounts of data. Graph-structured data often lacks the same scale as unstructured data, making the development of graph foundation models challenging. In this work, we propose Foundation-Informed Message Passing (FIMP), a Graph Neural Network (GNN) message-passing framework that leverages pretrained non-textual foundation models in graph-based tasks. We show that the self-attention layers of foundation models can effectively be repurposed on graphs to perform cross-node attention-based message-passing. Our model is evaluated on a real-world image network dataset and two biological applications (single-cell RNA sequencing data and fMRI brain activity recordings) in both finetuned and zero-shot settings. FIMP outperforms strong baselines, demonstrating that it can effectively leverage state-of-the-art foundation models in graph tasks.
- On the opportunities and risks of foundation models. arXiv preprint arXiv:2108.07258, 2021.
- Improving language understanding by generative pre-training. 2018.
- Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
- Uniter: Universal image-text representation learning. In European conference on computer vision, pages 104–120. Springer, 2020.
- Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
- Learning transferable visual models from natural language supervision. In International conference on machine learning, pages 8748–8763. PMLR, 2021.
- Florence: A new foundation model for computer vision. arXiv preprint arXiv:2111.11432, 2021.
- Don’t stop pretraining: Adapt language models to domains and tasks. arXiv preprint arXiv:2004.10964, 2020.
- scgpt: towards building a foundation model for single-cell multi-omics using generative ai. bioRxiv, pages 2023–04, 2023.
- Transfer learning enables predictions in network biology. Nature, pages 1–9, 2023.
- Brainlm: A foundation model for brain activity recordings. bioRxiv, pages 2023–09, 2023.
- Towards graph foundation models: A survey and beyond. arXiv preprint arXiv:2310.11829, 2023.
- Learning multiple layers of features from tiny images. 2009.
- Attention is all you need. Advances in neural information processing systems, 30, 2017.
- Language models are unsupervised multitask learners. OpenAI blog, 1(8):9, 2019.
- Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288, 2023.
- Emerging properties in self-supervised vision transformers. In Proceedings of the IEEE/CVF international conference on computer vision, pages 9650–9660, 2021.
- Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 9729–9738, 2020.
- Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF international conference on computer vision, pages 10012–10022, 2021.
- Graph contrastive learning with augmentations. Advances in neural information processing systems, 33:5812–5823, 2020.
- Deep graph contrastive representation learning. arXiv preprint arXiv:2006.04131, 2020.
- Deep graph infomax. ICLR (Poster), 2(3):4, 2019.
- Recipe for a general, powerful, scalable graph transformer. Advances in Neural Information Processing Systems, 35:14501–14515, 2022.
- Gpt4graph: Can large language models understand graph structured data? an empirical evaluation and benchmarking. arXiv preprint arXiv:2305.15066, 2023.
- Can language models solve graph problems in natural language? arXiv preprint arXiv:2305.10037, 2023.
- Multimodal population brain imaging in the uk biobank prospective epidemiological study. Nature neuroscience, 19(11):1523–1536, 2016.
- A unique brain connectome fingerprint predates and predicts response to antidepressants. IScience, 23(1), 2020.
- Highly sensitive spatial transcriptomics at near-cellular resolution with slide-seqv2. Nature biotechnology, 39(3):313–319, 2021.
- Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907, 2016.
- Inductive representation learning on large graphs. Advances in neural information processing systems, 30, 2017.
- Graph attention networks. arXiv preprint arXiv:1710.10903, 2017.
- How powerful are graph neural networks? arXiv preprint arXiv:1810.00826, 2018.
- Graphmae: Self-supervised masked graph autoencoders. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pages 594–604, 2022.
- Graphsaint: Graph sampling based inductive learning method. arXiv preprint arXiv:1907.04931, 2019.
- Fast graph representation learning with pytorch geometric. arXiv preprint arXiv:1903.02428, 2019.
- Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 16000–16009, 2022.
- Genept: A simple but hard-to-beat foundation model for genes and cells built from chatgpt. bioRxiv, pages 2023–10, 2023.
- An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020.
- Automatic denoising of functional mri data: combining independent component analysis and hierarchical fusion of classifiers. Neuroimage, 90:449–468, 2014.
- Chadi G Abdallah. Brain networks associated with covid-19 risk: Data from 3662 participants. Chronic Stress, 5:24705470211066770, 2021.
- Squidpy: a scalable framework for spatial omics analysis. Nature methods, 19(2):171–178, 2022.
- Collective classification in network data. AI magazine, 29(3):93–93, 2008.
- Pitfalls of graph neural network evaluation. arXiv preprint arXiv:1811.05868, 2018.
- Large-scale representation learning on graphs via bootstrapping. arXiv preprint arXiv:2102.06514, 2021.
- node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD international conference on Knowledge discovery and data mining, pages 855–864, 2016.
- The structural basis of age-related decline in global motion perception at fast and slow speeds. Neuropsychologia, 183:108507, 2023.
- The brain-body disconnect: A somatic sensory basis for trauma-related disorders. Frontiers in Neuroscience, 16:1881, 2022.
- Revisiting the role of cd63 as pro-tumorigenic or anti-tumorigenic tetraspanin in cancers and its theragnostic implications. Advanced Biology, page 2300078, 2023.
- Co-expression of timp-1 and its cell surface binding partner cd63 in glioblastomas. BMC cancer, 18:1–16, 2018.
- Microsoft academic graph: When experts are not enough. Quantitative Science Studies, 1(1):396–413, 2020.
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.