LLMs on Graphs
Introduction
Traditionally, LLMs have excelled in natural language processing by effectively encoding and decoding text. However, real-world data often presents itself in structured forms like graphs, which include academic and e-commerce networks or molecules paired with textual descriptions. While LLMs have showcased reasoning capabilities in text-based scenarios, their application in graph environments remains underexplored. By reviewing LLM techniques on graphs, this paper aims to present a systematic overview, categorizing potential scenarios and techniques, and suggesting future research directions.
Categories of Graph Scenarios
The application scenarios for LLMs on graphs are identified as:
- Pure Graphs involve data with no or minimal textual information, challenging LLMs' reasoning abilities on graph theory tasks.
- Text-Rich Graphs present scenarios where nodes or edges of a graph are associated with rich text, demanding models capable of understanding both text information and graph structures.
- Text-Paired Graphs feature graphs with overarching text descriptions, as seen in molecules with text notations, requiring models to jointly leverage molecular structures and associated textual information.
LLM Techniques on Graphs
Based on the role of LLMs, techniques on graphs are categorized as follows:
- LLM as Predictor includes methods transforming graphs to text sequences or employing LLM architectures that encode text and graph information concurrently. Techniques also involve LLM fine-tuning using graph structure supervision.
- LLM as Encoder describes using LLMs for initial text encoding before graph neural networks (GNNs) perform structure encoding. Challenges include convergence and sparsity issues, addressed through optimization, data augmentation, and knowledge distillation.
- LLM as Aligner relates to models aligning LLMs with GNNs for mutual enhancement, employing methods like prediction and latent space alignment.
Applications and Benchmarks
LLMs on graphs are applicable in domains such as scientific discovery and computational social science. They can also be used for virtual screening, optimizing scientific hypotheses, and synthesis planning. Benchmark datasets cover pure graphs, text-rich networks, and text-paired graphs. Evaluation metrics vary across tasks, often focusing on accuracy, precision, or domain-specific criteria.
Future Research Directions
Areas for future investigation encompass the creation of better benchmark datasets, exploration of broader task spaces with LLMs, design of multi-modal foundation models, efficient optimization of LLMs on graphs, and development of generalizable and robust LLMs for graph environments.
Conclusion
Despite strides in applying LLMs to graphs, significant challenges and questions remain. Addressing these can unlock the potential for LLMs to enhance our understanding across diverse graph scenarios, contributing to more advanced problem-solving and decision-making processes.