Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Improving Automatic Parallel Training via Balanced Memory Workload Optimization (2307.02031v2)

Published 5 Jul 2023 in cs.LG, cs.DB, and cs.DC
Improving Automatic Parallel Training via Balanced Memory Workload Optimization

Abstract: Transformer models have emerged as the leading approach for achieving state-of-the-art performance across various application domains, serving as the foundation for advanced large-scale deep learning (DL) models. However, efficiently training these models across multiple GPUs remains a complex challenge due to the abundance of parallelism options. Existing DL systems either require manual efforts to design distributed training plans or limit parallelism combinations to a constrained search space. In this paper, we present Galvatron-BMW, a novel system framework that integrates multiple prevalent parallelism dimensions and automatically identifies the most efficient hybrid parallelism strategy. To effectively navigate this vast search space, we employ a decision tree approach for decomposition and pruning based on intuitive insights. We further utilize a dynamic programming search algorithm to derive the optimal plan. Moreover, to improve resource utilization and enhance system efficiency, we propose a bi-objective optimization workflow that focuses on workload balance. Our evaluations on different Transformer models demonstrate the capabilities of Galvatron-BMW in automating distributed training under varying GPU memory constraints. Across all tested scenarios, Galvatron-BMW consistently achieves superior system throughput, surpassing previous approaches that rely on limited parallelism strategies.

Improving Automatic Parallel Training via Balanced Memory Workload Optimization

The paper "Improving Automatic Parallel Training via Balanced Memory Workload Optimization" introduces a novel system framework named Galvatron-BMW aimed at addressing the complex challenge of efficiently training Transformer models across multiple GPUs. Transformers have emerged as pivotal models in various domains, including NLP, CV, and recommendation systems, due to their state-of-the-art performance. However, their deployment is resource-intensive, necessitating intricate parallelism strategies to manage computational and memory requirements effectively.

Overview of Galvatron-BMW

Galvatron-BMW presents a comprehensive solution by automating the selection of efficient hybrid parallelism strategies for distributed training, leveraging a multi-faceted approach to parallelization. It incorporates multiple prevalent dimensions of parallelism — Data Parallelism (DP), Sharded Data Parallelism (SDP), Tensor Parallelism (TP), Pipeline Parallelism (PP), and Activation Checkpointing (CKPT) — addressing both memory consumption and computational overhead. These dimensions are automatically integrated through a decision tree approach for structured decomposition of the vast search space. This method is further refined by employing dynamic programming to derive optimal plans, coupling these with a bi-objective optimization workflow focusing on workload balance to enhance resource utilization and system efficiency.

Numerical Results and Framework Validation

The paper provides empirical evidence of Galvatron-BMW's efficacy through evaluations using various Transformer models under different GPU memory constraints. The results indicate that Galvatron-BMW consistently surpasses existing approaches, achieving significant throughput improvements. For instance, its advanced hybrid parallelism strategies enable it to accommodate larger batch sizes compared to other models while managing memory constraints more effectively. This accomplishment underscores the framework's capacity to enhance the overall system performance.

Implications and Future Directions

The implications of this research are substantial, both practically and theoretically. From a practical standpoint, Galvatron-BMW facilitates more efficient training of large-scale deep learning models, potentially reducing the time and cost associated with these processes. Theoretically, it contributes to the ongoing discourse on optimizing distributed training systems and may inspire future research into even more seamless integration of multiple parallelization dimensions using automatic, intelligent strategies.

Looking forward, potential advancements may include extending this framework to heterogeneous environments and addressing training efficiency for even larger models with dynamic, complex architectures. Additionally, exploring further automation in adjusting for variable hardware configurations could also significantly benefit the field.

In conclusion, the paper offers an insightful approach toward automatic hybrid parallelism in distributed deep learning, providing robust tools for improving throughput and efficiency in large-scale model training. Galvatron-BMW sets a precedent for future developments in this domain, promising enhancements in both the deployment and the theoretical understanding of parallel training strategies.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Yujie Wang (103 papers)
  2. Youhe Jiang (13 papers)
  3. Xupeng Miao (37 papers)
  4. Fangcheng Fu (31 papers)
  5. Xiaonan Nie (20 papers)
  6. Bin Cui (165 papers)
  7. Shenhan Zhu (5 papers)
  8. Yaofeng Tu (7 papers)
Citations (4)