Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Comprehensive Study on Large-Scale Graph Training: Benchmarking and Rethinking (2210.07494v2)

Published 14 Oct 2022 in cs.LG and cs.AI

Abstract: Large-scale graph training is a notoriously challenging problem for graph neural networks (GNNs). Due to the nature of evolving graph structures into the training process, vanilla GNNs usually fail to scale up, limited by the GPU memory space. Up to now, though numerous scalable GNN architectures have been proposed, we still lack a comprehensive survey and fair benchmark of this reservoir to find the rationale for designing scalable GNNs. To this end, we first systematically formulate the representative methods of large-scale graph training into several branches and further establish a fair and consistent benchmark for them by a greedy hyperparameter searching. In addition, regarding efficiency, we theoretically evaluate the time and space complexity of various branches and empirically compare them w.r.t GPU memory usage, throughput, and convergence. Furthermore, We analyze the pros and cons for various branches of scalable GNNs and then present a new ensembling training manner, named EnGCN, to address the existing issues. Our code is available at https://github.com/VITA-Group/Large_Scale_GCN_Benchmarking.

A Comprehensive Study on Large-Scale Graph Training: Benchmarking and Rethinking

This paper addresses the scalability challenges of large-scale graph training within the field of Graph Neural Networks (GNNs). Despite the rapid progression of scalable GNN architectures, a comprehensive survey and fair benchmark have been lacking to consolidate the rationale behind these designs. The authors aim to systematically formulate large-scale graph training methods, establish benchmarks, and propose improvements.

Overview of the Research

The primary focus of this paper is the scalability of GNNs, where traditional message passing approaches often consume prohibitive memory and computing resources. The paper categorizes existing scalable GNN methods into two main branches: Sampling-based methods and Decoupling-based methods. Sampling-based methods offer solutions to alleviate GPU memory usage by approximating full-batch training through various sampling strategies. Conversely, Decoupling-based methods separate message passing from feature transformation, potentially lowering the computational demand during training by utilizing CPU resources for preprocessing.

The benchmarks established in this paper assess the efficiency and effectiveness of these methods concerning accuracy, memory usage, throughput, and convergence. By performing an empirical analysis, optimal hyperparameter settings are identified and serve as a foundation for comparisons across methods. Notable datasets like Flickr, Reddit, and ogbn-products are employed, covering node counts from tens of thousands to millions.

Key Findings and Implications

The paper makes several non-trivial observations:

  1. Sensitivity to Hyperparameters: Sampling-based methods are noticeably sensitive to network depth and batch size, with performance correlating positively with batch size. This suggests that the underlying connectivity captured in training batches plays a critical role in model effectiveness.
  2. Performance of Precomputing-based Methods: These methods outperform others on larger datasets due to their stable and fast convergence rates. However, they are constrained by CPU memory requirements, which become particularly burdensome for exceedingly large graphs.
  3. Novel Training Frameworks: Alongside benchmarking, the authors introduce EnGCN, an ensembling-based training approach designed to mitigate existing scalability issues without exhaustive computational demands. This method achieves state-of-the-art results on various scales, primarily by sequentially training models across layers and applying ensemble strategies during inference.

Future Directions in AI Development

The research outlined by the authors posits several potential avenues for advancement. Incorporating self-label enhancements and sophisticated node feature extraction schemes could further elevate the performance of precomputing-based methods. Furthermore, hybrid models that blend sampling efficacy with ensembling strategies offer promising directions to efficiently handle large-scale graphs, potentially applicable in domains requiring real-time or resource-constrained inference.

As the horizon of AI continues to expand towards more complex and larger datasets, the implications of scalable GNNs are profound. Methods such as EnGCN could serve as foundational blocks for more intricate architectures, leading to substantial economic and computational benefits across industries where understanding large-scale interconnected data is paramount.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Keyu Duan (10 papers)
  2. Zirui Liu (58 papers)
  3. Peihao Wang (43 papers)
  4. Wenqing Zheng (16 papers)
  5. Kaixiong Zhou (52 papers)
  6. Tianlong Chen (202 papers)
  7. Xia Hu (186 papers)
  8. Zhangyang Wang (375 papers)
Citations (54)