GPT4Graph: Can Large Language Models Understand Graph Structured Data ? An Empirical Evaluation and Benchmarking (2305.15066v2)
Abstract: LLMs~(LLM) like ChatGPT have become indispensable to artificial general intelligence~(AGI), demonstrating excellent performance in various natural language processing tasks. In the real world, graph data is ubiquitous and an essential part of AGI and prevails in domains like social network analysis, bioinformatics and recommender systems. The training corpus of LLMs often includes some algorithmic components, which allows them to achieve certain effects on some graph data-related problems. However, there is still little research on their performance on a broader range of graph-structured data. In this study, we conduct an extensive investigation to assess the proficiency of LLMs in comprehending graph data, employing a diverse range of structural and semantic-related tasks. Our analysis encompasses 10 distinct tasks that evaluate the LLMs' capabilities in graph understanding. Through our study, we not only uncover the current limitations of LLMs in comprehending graph structures and performing associated reasoning tasks but also emphasize the necessity for further advancements and novel approaches to enhance their graph processing capabilities. Our findings contribute valuable insights towards bridging the gap between LLMs and graph understanding, paving the way for more effective graph mining and knowledge extraction.
- Marc Barthelemy. 2004. Betweenness centrality in large complex networks. The European physical journal B, 38(2):163–168.
- Bioinformatics. John Wiley & Sons.
- Node classification in social networks. arXiv preprint arXiv:1101.3291.
- Richard J Bolton and David J Hand. 2002. Statistical fraud detection: A review. Statistical science, 17(3):235–255.
- Ulrik Brandes. 2001. A faster algorithm for betweenness centrality. Journal of mathematical sociology, 25(2):163–177.
- Graph markup language (graphml).
- Towards table-to-text generation with pretrained language model: A table structure understanding and text deliberating approach. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 8199–8210, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
- An upper bound on the diameter of a graph from eigenvalues associated with its laplacian. SIAM Journal on Discrete Mathematics, 7(3):443–457.
- Mate: multi-view attention for table transformer efficiency. Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing".
- A fair comparison of graph neural networks for graph classification. arXiv preprint arXiv:1912.09893.
- Tablegpt: Few-shot table-to-text generation with table structure reconstruction and content matching. In Proceedings of the 28th International Conference on Computational Linguistics, pages 1978–1988.
- Michael Himsolt. 1997. Gml: Graph modelling language. University of Passau.
- Open graph benchmark: Datasets for machine learning on graphs. Advances in neural information processing systems, 33:22118–22133.
- Knowledge graph embedding based question answering. In Proceedings of the twelfth ACM international conference on web search and data mining, pages 105–113.
- Recommendation systems: Principles, methods and evaluation. Egyptian informatics journal, 16(3):261–273.
- A survey on knowledge graphs: Representation, acquisition, and applications. IEEE transactions on neural networks and learning systems, 33(2):494–514.
- Structgpt: A general framework for large language model to reason over structured data. arXiv preprint arXiv:2305.09645.
- Thomas N Kipf and Max Welling. 2016. Semi-supervised classification with graph convolutional networks. In International Conference on Learning Representations.
- Large language models are zero-shot reasoners. In Advances in Neural Information Processing Systems.
- PLOG: Table-to-logic pretraining for logical table-to-text generation. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 5531–5546, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
- Mark EJ Newman. 2005. A measure of betweenness centrality based on random walks. Social networks, 27(1):39–54.
- Ranking of closeness centrality for large-scale social networks. Lecture Notes in Computer Science, 5059:186–195.
- Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730–27744.
- Rethinking table recognition using graph neural networks. In 2019 International Conference on Document Analysis and Recognition (ICDAR), pages 142–147. IEEE.
- Dropedge: Towards deep graph convolutional networks on node classification. arXiv preprint arXiv:1907.10903.
- Toolformer: Language models can teach themselves to use tools. arXiv preprint arXiv:2302.04761.
- Evaluating and enhancing structural understanding capabilities of large language models on tables via input designs. arXiv preprint arXiv:2305.13062.
- Arnetminer: extraction and mining of academic social networks. In Proceedings of the 14th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 990–998.
- Graph attention networks. stat, 1050(20):10–48550.
- W Patrick Walters and Regina Barzilay. 2020. Applications of deep learning in molecule generation and molecular property prediction. Accounts of chemical research, 54(2):263–270.
- Stanley Wasserman and Katherine Faust. 1994. Social network analysis: Methods and applications.
- Finetuned language models are zero-shot learners. In International Conference on Learning Representations.
- Tree of thoughts: Deliberate problem solving with large language models. arXiv preprint arXiv:2305.10601.
- Jiawei Zhang. 2023. Graph-toolformer: To empower llms with graph reasoning ability via prompt augmented by chatgpt. arXiv preprint arXiv:2304.11116.
- Degree centrality, betweenness centrality, and closeness centrality in social network. In 2017 2nd international conference on modelling, simulation and applied mathematics (MSAM2017), pages 300–303. Atlantis press.
- Variational reasoning for question answering with knowledge graph. In Proceedings of the AAAI conference on artificial intelligence, volume 32.
- Data augmentation for graph neural networks. In Proceedings of the aaai conference on artificial intelligence, volume 35, pages 11015–11023.
- A survey of large language models. arXiv preprint arXiv:2303.18223.
- Jiayan Guo (15 papers)
- Lun Du (50 papers)
- Hengyu Liu (30 papers)
- Mengyu Zhou (24 papers)
- Xinyi He (19 papers)
- Shi Han (74 papers)