Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Data Agnostic Filter Gating for Efficient Deep Networks (2010.15041v1)

Published 28 Oct 2020 in cs.CV

Abstract: To deploy a well-trained CNN model on low-end computation edge devices, it is usually supposed to compress or prune the model under certain computation budget (e.g., FLOPs). Current filter pruning methods mainly leverage feature maps to generate important scores for filters and prune those with smaller scores, which ignores the variance of input batches to the difference in sparse structure over filters. In this paper, we propose a data agnostic filter pruning method that uses an auxiliary network named Dagger module to induce pruning and takes pretrained weights as input to learn the importance of each filter. In addition, to help prune filters with certain FLOPs constraints, we leverage an explicit FLOPs-aware regularization to directly promote pruning filters toward target FLOPs. Extensive experimental results on CIFAR-10 and ImageNet datasets indicate our superiority to other state-of-the-art filter pruning methods. For example, our 50\% FLOPs ResNet-50 can achieve 76.1\% Top-1 accuracy on ImageNet dataset, surpassing many other filter pruning methods.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Xiu Su (22 papers)
  2. Shan You (46 papers)
  3. Tao Huang (203 papers)
  4. Hongyan Xu (9 papers)
  5. Fei Wang (574 papers)
  6. Chen Qian (226 papers)
  7. Changshui Zhang (81 papers)
  8. Chang Xu (323 papers)
Citations (10)

Summary

We haven't generated a summary for this paper yet.