Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Efficient and effective training of language and graph neural network models (2206.10781v1)

Published 22 Jun 2022 in cs.LG and cs.CL

Abstract: Can we combine heterogenous graph structure with text to learn high-quality semantic and behavioural representations? Graph neural networks (GNN)s encode numerical node attributes and graph structure to achieve impressive performance in a variety of supervised learning tasks. Current GNN approaches are challenged by textual features, which typically need to be encoded to a numerical vector before provided to the GNN that may incur some information loss. In this paper, we put forth an efficient and effective framework termed LLM GNN (LM-GNN) to jointly train large-scale LLMs and graph neural networks. The effectiveness in our framework is achieved by applying stage-wise fine-tuning of the BERT model first with heterogenous graph information and then with a GNN model. Several system and design optimizations are proposed to enable scalable and efficient training. LM-GNN accommodates node and edge classification as well as link prediction tasks. We evaluate the LM-GNN framework in different datasets performance and showcase the effectiveness of the proposed approach. LM-GNN provides competitive results in an Amazon query-purchase-product application.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Vassilis N. Ioannidis (34 papers)
  2. Xiang Song (34 papers)
  3. Da Zheng (50 papers)
  4. Houyu Zhang (4 papers)
  5. Jun Ma (347 papers)
  6. Yi Xu (304 papers)
  7. Belinda Zeng (16 papers)
  8. Trishul Chilimbi (22 papers)
  9. George Karypis (110 papers)
Citations (9)

Summary

We haven't generated a summary for this paper yet.