Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Auto-MAP: A DQN Framework for Exploring Distributed Execution Plans for DNN Workloads (2007.04069v1)

Published 8 Jul 2020 in cs.DC and cs.AI

Abstract: The last decade has witnessed growth in the computational requirements for training deep neural networks. Current approaches (e.g., data/model parallelism, pipeline parallelism) parallelize training tasks onto multiple devices. However, these approaches always rely on specific deep learning frameworks and requires elaborate manual design, which make it difficult to maintain and share between different type of models. In this paper, we propose Auto-MAP, a framework for exploring distributed execution plans for DNN workloads, which can automatically discovering fast parallelization strategies through reinforcement learning on IR level of deep learning models. Efficient exploration remains a major challenge for reinforcement learning. We leverage DQN with task-specific pruning strategies to help efficiently explore the search space including optimized strategies. Our evaluation shows that Auto-MAP can find the optimal solution in two hours, while achieving better throughput on several NLP and convolution models.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Siyu Wang (55 papers)
  2. Yi Rong (12 papers)
  3. Shiqing Fan (10 papers)
  4. Zhen Zheng (39 papers)
  5. Guoping Long (12 papers)
  6. Jun Yang (357 papers)
  7. Xiaoyong Liu (6 papers)
  8. Wei Lin (207 papers)
  9. Lansong Diao (10 papers)
Citations (9)