Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Graph-Aware Language Model Pre-Training on a Large Graph Corpus Can Help Multiple Graph Applications (2306.02592v1)

Published 5 Jun 2023 in cs.CL, cs.AI, and cs.LG

Abstract: Model pre-training on large text corpora has been demonstrated effective for various downstream applications in the NLP domain. In the graph mining domain, a similar analogy can be drawn for pre-training graph models on large graphs in the hope of benefiting downstream graph applications, which has also been explored by several recent studies. However, no existing study has ever investigated the pre-training of text plus graph models on large heterogeneous graphs with abundant textual information (a.k.a. large graph corpora) and then fine-tuning the model on different related downstream applications with different graph schemas. To address this problem, we propose a framework of graph-aware LLM pre-training (GALM) on a large graph corpus, which incorporates LLMs and graph neural networks, and a variety of fine-tuning methods on downstream applications. We conduct extensive experiments on Amazon's real internal datasets and large public datasets. Comprehensive empirical results and in-depth analysis demonstrate the effectiveness of our proposed methods along with lessons learned.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (12)
  1. Han Xie (21 papers)
  2. Da Zheng (50 papers)
  3. Jun Ma (347 papers)
  4. Houyu Zhang (4 papers)
  5. Vassilis N. Ioannidis (34 papers)
  6. Xiang Song (34 papers)
  7. Qing Ping (13 papers)
  8. Sheng Wang (239 papers)
  9. Carl Yang (130 papers)
  10. Yi Xu (302 papers)
  11. Belinda Zeng (16 papers)
  12. Trishul Chilimbi (22 papers)
Citations (24)