Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

PruneTrain: Fast Neural Network Training by Dynamic Sparse Model Reconfiguration (1901.09290v5)

Published 26 Jan 2019 in cs.LG and stat.ML

Abstract: State-of-the-art convolutional neural networks (CNNs) used in vision applications have large models with numerous weights. Training these models is very compute- and memory-resource intensive. Much research has been done on pruning or compressing these models to reduce the cost of inference, but little work has addressed the costs of training. We focus precisely on accelerating training. We propose PruneTrain, a cost-efficient mechanism that gradually reduces the training cost during training. PruneTrain uses a structured group-lasso regularization approach that drives the training optimization toward both high accuracy and small weight values. Small weights can then be periodically removed by reconfiguring the network model to a smaller one. By using a structured-pruning approach and additional reconfiguration techniques we introduce, the pruned model can still be efficiently processed on a GPU accelerator. Overall, PruneTrain achieves a reduction of 39% in the end-to-end training time of ResNet50 for ImageNet by reducing computation cost by 40% in FLOPs, memory accesses by 37% for memory bandwidth bound layers, and the inter-accelerator communication by 55%.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Sangkug Lym (7 papers)
  2. Esha Choukse (15 papers)
  3. Siavash Zangeneh (1 paper)
  4. Wei Wen (49 papers)
  5. Sujay Sanghavi (97 papers)
  6. Mattan Erez (16 papers)
Citations (85)

Summary

We haven't generated a summary for this paper yet.