Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Conditional Automated Channel Pruning for Deep Neural Networks (2009.09724v2)

Published 21 Sep 2020 in cs.CV

Abstract: Model compression aims to reduce the redundancy of deep networks to obtain compact models. Recently, channel pruning has become one of the predominant compression methods to deploy deep models on resource-constrained devices. Most channel pruning methods often use a fixed compression rate for all the layers of the model, which, however, may not be optimal. To address this issue, given a target compression rate for the whole model, one can search for the optimal compression rate for each layer. Nevertheless, these methods perform channel pruning for a specific target compression rate. When we consider multiple compression rates, they have to repeat the channel pruning process multiple times, which is very inefficient yet unnecessary. To address this issue, we propose a Conditional Automated Channel Pruning(CACP) method to obtain the compressed models with different compression rates through single channel pruning process. To this end, we develop a conditional model that takes an arbitrary compression rate as input and outputs the corresponding compressed model. In the experiments, the resultant models with different compression rates consistently outperform the models compressed by existing methods with a channel pruning process for each target compression rate.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Yixin Liu (108 papers)
  2. Yong Guo (67 papers)
  3. Zichang Liu (14 papers)
  4. Haohua Liu (1 paper)
  5. Jingjie Zhang (13 papers)
  6. Zejun Chen (1 paper)
  7. Jing Liu (526 papers)
  8. Jian Chen (257 papers)
Citations (15)

Summary

We haven't generated a summary for this paper yet.