Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 98 tok/s
Gemini 2.5 Pro 49 tok/s Pro
GPT-5 Medium 15 tok/s
GPT-5 High 16 tok/s Pro
GPT-4o 86 tok/s
GPT OSS 120B 470 tok/s Pro
Kimi K2 158 tok/s Pro
2000 character limit reached

FIMP: Foundation Model-Informed Message Passing for Graph Neural Networks (2210.09475v5)

Published 17 Oct 2022 in cs.LG

Abstract: Foundation models have achieved remarkable success across many domains, relying on pretraining over vast amounts of data. Graph-structured data often lacks the same scale as unstructured data, making the development of graph foundation models challenging. In this work, we propose Foundation-Informed Message Passing (FIMP), a Graph Neural Network (GNN) message-passing framework that leverages pretrained non-textual foundation models in graph-based tasks. We show that the self-attention layers of foundation models can effectively be repurposed on graphs to perform cross-node attention-based message-passing. Our model is evaluated on a real-world image network dataset and two biological applications (single-cell RNA sequencing data and fMRI brain activity recordings) in both finetuned and zero-shot settings. FIMP outperforms strong baselines, demonstrating that it can effectively leverage state-of-the-art foundation models in graph tasks.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (50)
  1. On the opportunities and risks of foundation models. arXiv preprint arXiv:2108.07258, 2021.
  2. Improving language understanding by generative pre-training. 2018.
  3. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
  4. Uniter: Universal image-text representation learning. In European conference on computer vision, pages 104–120. Springer, 2020.
  5. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
  6. Learning transferable visual models from natural language supervision. In International conference on machine learning, pages 8748–8763. PMLR, 2021.
  7. Florence: A new foundation model for computer vision. arXiv preprint arXiv:2111.11432, 2021.
  8. Don’t stop pretraining: Adapt language models to domains and tasks. arXiv preprint arXiv:2004.10964, 2020.
  9. scgpt: towards building a foundation model for single-cell multi-omics using generative ai. bioRxiv, pages 2023–04, 2023.
  10. Transfer learning enables predictions in network biology. Nature, pages 1–9, 2023.
  11. Brainlm: A foundation model for brain activity recordings. bioRxiv, pages 2023–09, 2023.
  12. Towards graph foundation models: A survey and beyond. arXiv preprint arXiv:2310.11829, 2023.
  13. Learning multiple layers of features from tiny images. 2009.
  14. Attention is all you need. Advances in neural information processing systems, 30, 2017.
  15. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9, 2019.
  16. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288, 2023.
  17. Emerging properties in self-supervised vision transformers. In Proceedings of the IEEE/CVF international conference on computer vision, pages 9650–9660, 2021.
  18. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 9729–9738, 2020.
  19. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF international conference on computer vision, pages 10012–10022, 2021.
  20. Graph contrastive learning with augmentations. Advances in neural information processing systems, 33:5812–5823, 2020.
  21. Deep graph contrastive representation learning. arXiv preprint arXiv:2006.04131, 2020.
  22. Deep graph infomax. ICLR (Poster), 2(3):4, 2019.
  23. Recipe for a general, powerful, scalable graph transformer. Advances in Neural Information Processing Systems, 35:14501–14515, 2022.
  24. Gpt4graph: Can large language models understand graph structured data? an empirical evaluation and benchmarking. arXiv preprint arXiv:2305.15066, 2023.
  25. Can language models solve graph problems in natural language? arXiv preprint arXiv:2305.10037, 2023.
  26. Multimodal population brain imaging in the uk biobank prospective epidemiological study. Nature neuroscience, 19(11):1523–1536, 2016.
  27. A unique brain connectome fingerprint predates and predicts response to antidepressants. IScience, 23(1), 2020.
  28. Highly sensitive spatial transcriptomics at near-cellular resolution with slide-seqv2. Nature biotechnology, 39(3):313–319, 2021.
  29. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907, 2016.
  30. Inductive representation learning on large graphs. Advances in neural information processing systems, 30, 2017.
  31. Graph attention networks. arXiv preprint arXiv:1710.10903, 2017.
  32. How powerful are graph neural networks? arXiv preprint arXiv:1810.00826, 2018.
  33. Graphmae: Self-supervised masked graph autoencoders. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pages 594–604, 2022.
  34. Graphsaint: Graph sampling based inductive learning method. arXiv preprint arXiv:1907.04931, 2019.
  35. Fast graph representation learning with pytorch geometric. arXiv preprint arXiv:1903.02428, 2019.
  36. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 16000–16009, 2022.
  37. Genept: A simple but hard-to-beat foundation model for genes and cells built from chatgpt. bioRxiv, pages 2023–10, 2023.
  38. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020.
  39. Automatic denoising of functional mri data: combining independent component analysis and hierarchical fusion of classifiers. Neuroimage, 90:449–468, 2014.
  40. Chadi G Abdallah. Brain networks associated with covid-19 risk: Data from 3662 participants. Chronic Stress, 5:24705470211066770, 2021.
  41. Squidpy: a scalable framework for spatial omics analysis. Nature methods, 19(2):171–178, 2022.
  42. Collective classification in network data. AI magazine, 29(3):93–93, 2008.
  43. Pitfalls of graph neural network evaluation. arXiv preprint arXiv:1811.05868, 2018.
  44. Large-scale representation learning on graphs via bootstrapping. arXiv preprint arXiv:2102.06514, 2021.
  45. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD international conference on Knowledge discovery and data mining, pages 855–864, 2016.
  46. The structural basis of age-related decline in global motion perception at fast and slow speeds. Neuropsychologia, 183:108507, 2023.
  47. The brain-body disconnect: A somatic sensory basis for trauma-related disorders. Frontiers in Neuroscience, 16:1881, 2022.
  48. Revisiting the role of cd63 as pro-tumorigenic or anti-tumorigenic tetraspanin in cancers and its theragnostic implications. Advanced Biology, page 2300078, 2023.
  49. Co-expression of timp-1 and its cell surface binding partner cd63 in glioblastomas. BMC cancer, 18:1–16, 2018.
  50. Microsoft academic graph: When experts are not enough. Quantitative Science Studies, 1(1):396–413, 2020.
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

  • The paper introduces the FIMP framework that uses pre-trained foundation model weights as message passing operators in GNNs.
  • It employs data-specific node tokenization and cross-node attention to enhance feature representation in graph-structured data.
  • Experimental results show performance gains of up to 25.8% in fMRI reconstruction and 17% in image graph tasks compared to baselines.

Review of "FIMP: Foundation Model-Informed Message Passing for Graph Neural Networks"

The paper "FIMP: Foundation Model-Informed Message Passing for Graph Neural Networks" introduces a novel framework integrating foundational model capabilities into Graph Neural Networks (GNNs). The Foundation-Informed Message Passing (FIMP) framework addresses a significant gap in leveraging large-scale pre-trained model weights—typically utilized in non-graph domains—within graph-structured settings, enhancing the applicability and performance of GNNs across various domains.

Summary of Contributions

The paper's primary contribution is the development of the FIMP framework, which adapts the weights from transformer-based foundation models as message-passing operators in GNN architectures. This method allows for the utilization of prior knowledge captured by LLMs and other foundational models during pretraining. The adaptation process involves the construction of message-passing operations directly informed by the transformed model weights, applying them to graph-structured data.

Key technical elements of this framework include:

  • Node Tokenization: FIMP uses data-specific tokenization schemes to encode node features into sequences suitable for processing by transformer models. This adaptation allows the framework to integrate smoothly with foundation models and effectively utilize their pre-trained weights in new domains.
  • Cross-Node Attention: The framework employs cross-node multihead attention, extending beyond traditional self-attention methods by using the destination node's features as the query sequence and the source node's features as the key and value sequences. This approach enables sophisticated message passing between nodes.
  • Adaptation to Pretrained Weights: By replacing the native GNN message-passing operations with pre-trained weights from foundation models, the framework enhances the graph's representation quality by incorporating extensive knowledge from prior large-scale unsupervised learning.

Experimental Results and Analysis

The paper evaluates the FIMP framework across multiple domains, including fMRI brain activity reconstruction, gene expression prediction in spatial transcriptomics datasets, and image graph construction from the CIFAR-100 dataset:

  • fMRI Brain Activity: The FIMP framework significantly improved the reconstruction accuracy by 25.8% over the nearest baseline, further enhancing performance by utilizing foundation models trained on brain activity like BrainLM.
  • Spatial Transcriptomics: FIMP consistently outperformed existing baselines on masked gene expression prediction tasks. The use of pretrained models in limited data settings resulted in further performance gains, illustrating FIMP’s potential in cross-domain applications.
  • Image Graph Data: On CIFAR-100, FIMP demonstrated a 17% improvement in reconstruction accuracy compared to traditional GNN approaches, underscoring the utility of integrating foundation model vision capabilities.

These results demonstrate that FIMP can exploit sophisticated representation learned during large-scale pretraining to tackle diverse graph-based tasks effectively. The paper's contributions are especially pertinent in fields relying on complex feature-rich data, where traditional GNNs may struggle to capture nuanced inter-feature relationships.

Implications and Future Directions

The introduction of FIMP opens avenues for integrating domain-agnostic pre-trained models with domain-specific graph tasks, a strategy that could bridge knowledge between distinct modalities. This advance implies a better transfer of insights across disparate areas, does not require bespoke model design for each task, and ensures robust performance despite limited graph structure or node feature specificity.

Potential future developments include enhancing token selection strategies for sparse datasets, integrating efficient attention mechanisms for computational scalability, and incorporating richer context from edge features. An interesting direction might also be the exploration of joint training techniques, leveraging foundation model knowledge while emphasizing specific graph-based task requirements.

In conclusion, this paper presents a compelling approach to expanding the applicability of foundation models beyond traditional domains, providing a robust framework for improving graph neural network capabilities.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Paper Prompts

Sign up for free to create and run prompts on this paper using GPT-5.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com