Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Towards Compact CNNs via Collaborative Compression (2105.11228v1)

Published 24 May 2021 in cs.CV and cs.AI

Abstract: Channel pruning and tensor decomposition have received extensive attention in convolutional neural network compression. However, these two techniques are traditionally deployed in an isolated manner, leading to significant accuracy drop when pursuing high compression rates. In this paper, we propose a Collaborative Compression (CC) scheme, which joints channel pruning and tensor decomposition to compress CNN models by simultaneously learning the model sparsity and low-rankness. Specifically, we first investigate the compression sensitivity of each layer in the network, and then propose a Global Compression Rate Optimization that transforms the decision problem of compression rate into an optimization problem. After that, we propose multi-step heuristic compression to remove redundant compression units step-by-step, which fully considers the effect of the remaining compression space (i.e., unremoved compression units). Our method demonstrates superior performance gains over previous ones on various datasets and backbone architectures. For example, we achieve 52.9% FLOPs reduction by removing 48.4% parameters on ResNet-50 with only a Top-1 accuracy drop of 0.56% on ImageNet 2012.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Yuchao Li (24 papers)
  2. Shaohui Lin (45 papers)
  3. Jianzhuang Liu (91 papers)
  4. Qixiang Ye (110 papers)
  5. Mengdi Wang (199 papers)
  6. Fei Chao (53 papers)
  7. Fan Yang (878 papers)
  8. Jincheng Ma (1 paper)
  9. Qi Tian (314 papers)
  10. Rongrong Ji (315 papers)
Citations (78)

Summary

We haven't generated a summary for this paper yet.