Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Dorylus: Affordable, Scalable, and Accurate GNN Training with Distributed CPU Servers and Serverless Threads (2105.11118v2)

Published 24 May 2021 in cs.DC and cs.LG

Abstract: A graph neural network (GNN) enables deep learning on structured graph data. There are two major GNN training obstacles: 1) it relies on high-end servers with many GPUs which are expensive to purchase and maintain, and 2) limited memory on GPUs cannot scale to today's billion-edge graphs. This paper presents Dorylus: a distributed system for training GNNs. Uniquely, Dorylus can take advantage of serverless computing to increase scalability at a low cost. The key insight guiding our design is computation separation. Computation separation makes it possible to construct a deep, bounded-asynchronous pipeline where graph and tensor parallel tasks can fully overlap, effectively hiding the network latency incurred by Lambdas. With the help of thousands of Lambda threads, Dorylus scales GNN training to billion-edge graphs. Currently, for large graphs, CPU servers offer the best performance-per-dollar over GPU servers. Just using Lambdas on top of CPU servers offers up to 2.75x more performance-per-dollar than training only with CPU servers. Concretely, Dorylus is 1.22x faster and 4.83x cheaper than GPU servers for massive sparse graphs. Dorylus is up to 3.8x faster and 10.7x cheaper compared to existing sampling-based systems.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. John Thorpe (2 papers)
  2. Yifan Qiao (19 papers)
  3. Jonathan Eyolfson (2 papers)
  4. Shen Teng (1 paper)
  5. Guanzhou Hu (3 papers)
  6. Zhihao Jia (43 papers)
  7. Jinliang Wei (9 papers)
  8. Keval Vora (5 papers)
  9. Ravi Netravali (22 papers)
  10. Miryung Kim (17 papers)
  11. Guoqing Harry Xu (7 papers)
Citations (123)

Summary

We haven't generated a summary for this paper yet.