Inductive Graph Alignment Prompt: Bridging the Gap between Graph Pre-training and Inductive Fine-tuning From Spectral Perspective (2402.13556v1)
Abstract: The "Graph pre-training and fine-tuning" paradigm has significantly improved Graph Neural Networks(GNNs) by capturing general knowledge without manual annotations for downstream tasks. However, due to the immense gap of data and tasks between the pre-training and fine-tuning stages, the model performance is still limited. Inspired by prompt fine-tuning in Natural Language Processing(NLP), many endeavors have been made to bridge the gap in graph domain. But existing methods simply reformulate the form of fine-tuning tasks to the pre-training ones. With the premise that the pre-training graphs are compatible with the fine-tuning ones, these methods typically operate in transductive setting. In order to generalize graph pre-training to inductive scenario where the fine-tuning graphs might significantly differ from pre-training ones, we propose a novel graph prompt based method called Inductive Graph Alignment Prompt(IGAP). Firstly, we unify the mainstream graph pre-training frameworks and analyze the essence of graph pre-training from graph spectral theory. Then we identify the two sources of the data gap in inductive setting: (i) graph signal gap and (ii) graph structure gap. Based on the insight of graph pre-training, we propose to bridge the graph signal gap and the graph structure gap with learnable prompts in the spectral space. A theoretical analysis ensures the effectiveness of our method. At last, we conduct extensive experiments among nodes classification and graph classification tasks under the transductive, semi-inductive and inductive settings. The results demonstrate that our proposed method can successfully bridge the data gap under different settings.
- The pushshift reddit dataset. In Proceedings of the international AAAI conference on web and social media, Vol. 14. 830–839.
- Aleksandar Bojchevski and Stephan Günnemann. 2017. Deep gaussian embedding of graphs: Unsupervised inductive learning via ranking. arXiv preprint arXiv:1707.03815 (2017).
- Flex: Unifying evaluation for few-shot nlp. Advances in Neural Information Processing Systems 34 (2021), 15787–15800.
- Spectral networks and locally connected networks on graphs. In Proceedings of the 2nd International Conference on Learning Representations.
- A simple framework for contrastive learning of visual representations. In International conference on machine learning. PMLR, 1597–1607.
- Xinlei Chen and Kaiming He. 2021. Exploring simple siamese representation learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 15750–15758.
- Masked spectrogram prediction for self-supervised audio pre-training. In ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 1–5.
- Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018).
- Unified language model pre-training for natural language understanding and generation. Advances in neural information processing systems 32 (2019).
- Universal Prompt Tuning for Graph Neural Networks. arXiv preprint arXiv:2209.15240 (2022).
- Spatial-temporal graph ode networks for traffic flow forecasting. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining. 364–373.
- Bootstrap your own latent-a new approach to self-supervised learning. Advances in neural information processing systems 33 (2020), 21271–21284.
- Inductive Representation Learning on Large Graphs. In Proceedings of the 31st Conference on Neural Information Processing Systems.
- Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 9729–9738.
- Gpt-gnn: Generative pre-training of graph neural networks. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. 1857–1867.
- Visual prompt tuning. In European Conference on Computer Vision. Springer, 709–727.
- Scaling laws for neural language models. arXiv preprint arXiv:2001.08361 (2020).
- Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014).
- Thomas N. Kipf and Max Welling. 2017. Semi-Supervised Classification with Graph Convolutional Networks. In Proceedings of the 5th International Conference on Learning Representations.
- Greg Landrum. 2006. RDKit: Open-source cheminformatics. 2006. Google Scholar (2006).
- The power of scale for parameter-efficient prompt tuning. arXiv preprint arXiv:2104.08691 (2021).
- Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. Comput. Surveys 55, 9 (2023), 1–35.
- Vivian Liu and Lydia B Chilton. 2022. Design guidelines for prompt engineering text-to-image generative models. In Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems. 1–23.
- Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019).
- Graphprompt: Unifying pre-training and downstream tasks for graph neural networks. In Proceedings of the ACM Web Conference 2023. 417–428.
- Graph structural-topic neural network. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. 1065–1073.
- Theoretically improving graph neural networks via anonymous walk graph kernels. In Proceedings of the Web Conference 2021. 1204–1214.
- HGK-GNN: Heterogeneous Graph Kernel based Graph Neural Networks. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining. 1129–1138.
- Gcan: Graph convolutional adversarial network for unsupervised domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 8266–8276.
- Spatio-temporal deep graph infomax. arXiv preprint arXiv:1904.06316 (2019).
- Graph Representation Learning via Graphical Mutual Information Maximization. In Proceedings of The Web Conference 2020. 259–270.
- Learning transferable visual models from natural language supervision. In International conference on machine learning. PMLR, 8748–8763.
- Improving language understanding by generative pre-training. (2018).
- Curriculum graph co-teaching for multi-target domain adaptation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 5351–5360.
- Pitfalls of graph neural network evaluation. arXiv preprint arXiv:1811.05868 (2018).
- Gppt: Graph pre-training and prompt tuning to generalize graph neural networks. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 1717–1727.
- All in One: Multi-Task Prompting for Graph Neural Networks. (2023).
- Graph attention networks. In Proceedings of the 6th International Conference on Learning Representations.
- Deep Graph Infomax.. In ICLR (Poster).
- Microsoft academic graph: When experts are not enough. Quantitative Science Studies 1, 1 (2020), 396–413.
- Finetuned language models are zero-shot learners. arXiv preprint arXiv:2109.01652 (2021).
- Unsupervised domain adaptive graph convolutional networks. In Proceedings of The Web Conference 2020. 1457–1467.
- MoleculeNet: a benchmark for molecular machine learning. Chemical science 9, 2 (2018), 513–530.
- How Powerful are Graph Neural Networks?. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net.
- Jiaqi Zeng and Pengtao Xie. 2021. Contrastive self-supervised learning for graph classification. In Proceedings of the AAAI conference on Artificial Intelligence, Vol. 35. 10824–10832.
- Continual Learning on Dynamic Graphs via Parameter Isolation. arXiv preprint arXiv:2305.13825 (2023).
- Dane: Domain adaptive network embedding. arXiv preprint arXiv:1906.00684 (2019).
- Exploring recommendation capabilities of gpt-4v (ision): A preliminary case study. arXiv preprint arXiv:2311.04199 (2023).
- Deep graph contrastive representation learning. arXiv preprint arXiv:2006.04131 (2020).
- Graph contrastive learning with adaptive augmentation. In Proceedings of the Web Conference 2021. 2069–2080.
- Yuchen Yan (44 papers)
- Peiyan Zhang (21 papers)
- Zheng Fang (103 papers)
- Qingqing Long (25 papers)