Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

An Oracle for Guiding Large-Scale Model/Hybrid Parallel Training of Convolutional Neural Networks (2104.09075v1)

Published 19 Apr 2021 in cs.DC and cs.LG

Abstract: Deep Neural Network (DNN) frameworks use distributed training to enable faster time to convergence and alleviate memory capacity limitations when training large models and/or using high dimension inputs. With the steady increase in datasets and model sizes, model/hybrid parallelism is deemed to have an important role in the future of distributed training of DNNs. We analyze the compute, communication, and memory requirements of Convolutional Neural Networks (CNNs) to understand the trade-offs between different parallelism approaches on performance and scalability. We leverage our model-driven analysis to be the basis for an oracle utility which can help in detecting the limitations and bottlenecks of different parallelism approaches at scale. We evaluate the oracle on six parallelization strategies, with four CNN models and multiple datasets (2D and 3D), on up to 1024 GPUs. The results demonstrate that the oracle has an average accuracy of about 86.74% when compared to empirical results, and as high as 97.57% for data parallelism.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Albert Njoroge Kahira (4 papers)
  2. Truong Thao Nguyen (10 papers)
  3. Leonardo Bautista Gomez (5 papers)
  4. Ryousei Takano (16 papers)
  5. Rosa M Badia (35 papers)
  6. Mohamed Wahib (38 papers)
Citations (8)