HiGPT: Heterogeneous Graph Language Model (2402.16024v2)
Abstract: Heterogeneous graph learning aims to capture complex relationships and diverse relational semantics among entities in a heterogeneous graph to obtain meaningful representations for nodes and edges. Recent advancements in heterogeneous graph neural networks (HGNNs) have achieved state-of-the-art performance by considering relation heterogeneity and using specialized message functions and aggregation rules. However, existing frameworks for heterogeneous graph learning have limitations in generalizing across diverse heterogeneous graph datasets. Most of these frameworks follow the "pre-train" and "fine-tune" paradigm on the same dataset, which restricts their capacity to adapt to new and unseen data. This raises the question: "Can we generalize heterogeneous graph models to be well-adapted to diverse downstream learning tasks with distribution shifts in both node token sets and relation type heterogeneity?'' To tackle those challenges, we propose HiGPT, a general large graph model with Heterogeneous graph instruction-tuning paradigm. Our framework enables learning from arbitrary heterogeneous graphs without the need for any fine-tuning process from downstream datasets. To handle distribution shifts in heterogeneity, we introduce an in-context heterogeneous graph tokenizer that captures semantic relationships in different heterogeneous graphs, facilitating model adaptation. We incorporate a large corpus of heterogeneity-aware graph instructions into our HiGPT, enabling the model to effectively comprehend complex relation heterogeneity and distinguish between various types of graph tokens. Furthermore, we introduce the Mixture-of-Thought (MoT) instruction augmentation paradigm to mitigate data scarcity by generating diverse and informative instructions. Through comprehensive evaluations, our proposed framework demonstrates exceptional performance in terms of generalization performance.
- Graphllm: Boosting graph reasoning ability of large language model. arXiv preprint arXiv:2310.05845, 2023.
- Heterogeneous graph contrastive learning for recommendation. In WSDM, pages 544–552, 2023.
- Exploring the potential of large language models (llms) in learning on graphs. CoRR, abs/2307.03393, 2023.
- metapath2vec: Scalable representation learning for heterogeneous networks. In KDD, pages 135–144, 2017.
- Twhin: Embedding the twitter heterogeneous information network for personalized recommendation. In KDD, pages 2842–2850, 2022.
- Metapath-guided heterogeneous graph neural network for intent recommendation. In KDD, pages 2478–2486, 2019.
- Talk like a graph: Encoding graphs for large language models. arXiv preprint arXiv:2310.04560, 2023.
- MAGNN: metapath aggregated graph neural network for heterogeneous graph embedding. In WWW, pages 2331–2341. ACM / IW3C2, 2020.
- Inductive representation learning on large graphs. In NeurIPS, pages 1024–1034, 2017.
- Heterogeneous graph transformer. In WWW, pages 2704–2710. ACM / IW3C2, 2020.
- Self-supervised auxiliary learning with meta-paths for heterogeneous graphs. NeurIPS, 33:10294–10305, 2020.
- Heterogeneous graph neural network via attribute completion. In WWW, pages 391–400, 2021.
- X-goal: multiplex heterogeneous graph prototypical contrastive learning. In CIKM, pages 894–904, 2022.
- Heterogeneous graph attention networks for semi-supervised short text classification. In EMNLP, pages 4821–4830, 2019.
- Generated knowledge prompting for commonsense reasoning. In ACL (1), pages 3154–3169. Association for Computational Linguistics, 2022.
- Are we really making much progress? revisiting, benchmarking and refining heterogeneous graph neural networks. In KDD, pages 1150–1160, 2021.
- Single-cell biological network inference using a heterogeneous graph transformer. Nature Communications, 14(1):964, 2023.
- Rethinking the role of demonstrations: What makes in-context learning work? In EMNLP, pages 11048–11064, 2022.
- Unsupervised attributed multiplex network embedding. In AAAI, volume 34, pages 5371–5378, 2020.
- Learning transferable visual models from natural language supervision. In International Conference on Machine Learning (ICML), pages 8748–8763. PMLR, 2021.
- Hetegcn: heterogeneous graph convolutional networks for text classification. In WSDM, pages 860–868, 2021.
- N. Reimers and I. Gurevych. Sentence-bert: Sentence embeddings using siamese bert-networks. In EMNLP. Association for Computational Linguistics, 11 2019.
- Representation learning with large language models for recommendation. arXiv preprint arXiv:2310.15950, 2023.
- Distilling reasoning capabilities into smaller language models. In ACL, pages 7059–7073, 2023.
- Query-dependent prompt evaluation and optimization with offline inverse rl. arXiv e-prints, pages arXiv–2309, 2023.
- Graphgpt: Graph instruction tuning for large language models, 2023.
- Heterogeneous graph masked autoencoders. In AAAI, volume 37, pages 9997–10005, 2023.
- Graph attention networks. In ICLR (Poster). OpenReview.net, 2018.
- Relational message passing for knowledge graph completion. In KDD, pages 1697–1707, 2021.
- Self-supervised learning of contextual embeddings for link prediction in heterogeneous networks. In WWW, pages 2946–2957, 2021.
- A survey on heterogeneous graph embedding: methods, techniques, applications and sources. Transactions on Big Data (TBD), 9(2):415–436, 2022.
- Heterogeneous graph attention network. In WWW, pages 2022–2032, 2019.
- Heterogeneous graph attention network. In WWW, pages 2022–2032. ACM, 2019.
- Self-supervised heterogeneous graph neural network with co-contrastive learning. In KDD, pages 1726–1736, 2021.
- Chain-of-thought prompting elicits reasoning in large language models. In NeurIPS, 2022.
- Contrastive meta learning with behavior multiplicity for recommendation. In WSDM, pages 1120–1128, 2022.
- Llmrec: Large language models with graph augmentation for recommendation. arXiv preprint arXiv:2311.00423, 2023.
- Z. Wen and Y. Fang. Augmenting low-resource text classification with graph-grounded pre-training and prompting. In SIGIR, 2023.
- Knowledge enhancement for contrastive multi-behavior recommendation. In WSDM, pages 195–203, 2023.
- Heterogeneous network representation learning: A unified framework with survey and benchmark. Transactions on Knowledge and Data Engineering (TKDE), 34(10):4854–4873, 2020.
- Self-supervised heterogeneous graph pre-training based on structural clustering. NeurIPS, 35:16962–16974, 2022.
- Tree of thoughts: Deliberate problem solving with large language models. CoRR, abs/2305.10601, 2023.
- Natural language is all a graph needs. arXiv preprint arXiv:2308.07134, 2023.
- Heterogeneous graph neural network. In KDD, pages 793–803. ACM, 2019.
- Heterogeneous graph structure learning for graph neural networks. In AAAI, volume 35, pages 4697–4705, 2021.
- Jiabin Tang (15 papers)
- Yuhao Yang (23 papers)
- Wei Wei (424 papers)
- Lei Shi (262 papers)
- Long Xia (25 papers)
- Dawei Yin (165 papers)
- Chao Huang (244 papers)