Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Characterizing and Understanding Distributed GNN Training on GPUs (2204.08150v1)

Published 18 Apr 2022 in cs.DC and cs.LG

Abstract: Graph neural network (GNN) has been demonstrated to be a powerful model in many domains for its effectiveness in learning over graphs. To scale GNN training for large graphs, a widely adopted approach is distributed training which accelerates training using multiple computing nodes. Maximizing the performance is essential, but the execution of distributed GNN training remains preliminarily understood. In this work, we provide an in-depth analysis of distributed GNN training on GPUs, revealing several significant observations and providing useful guidelines for both software optimization and hardware optimization.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Haiyang Lin (3 papers)
  2. Mingyu Yan (39 papers)
  3. Xiaocheng Yang (11 papers)
  4. Mo Zou (4 papers)
  5. Wenming Li (12 papers)
  6. Xiaochun Ye (37 papers)
  7. Dongrui Fan (36 papers)
Citations (10)