Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Distilling Large Language Models for Text-Attributed Graph Learning (2402.12022v2)

Published 19 Feb 2024 in cs.CL and cs.LG

Abstract: Text-Attributed Graphs (TAGs) are graphs of connected textual documents. Graph models can efficiently learn TAGs, but their training heavily relies on human-annotated labels, which are scarce or even unavailable in many applications. LLMs have recently demonstrated remarkable capabilities in few-shot and zero-shot TAG learning, but they suffer from scalability, cost, and privacy issues. Therefore, in this work, we focus on synergizing LLMs and graph models with their complementary strengths by distilling the power of LLMs to a local graph model on TAG learning. To address the inherent gaps between LLMs (generative models for texts) and graph models (discriminative models for graphs), we propose first to let LLMs teach an interpreter with rich textual rationale and then let a student model mimic the interpreter's reasoning without LLMs' textual rationale. Extensive experiments validate the efficacy of our proposed framework.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Bo Pan (31 papers)
  2. Zheng Zhang (486 papers)
  3. Yifei Zhang (167 papers)
  4. Yuntong Hu (12 papers)
  5. Liang Zhao (353 papers)
Citations (5)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets