Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Channel Pruning Guided by Spatial and Channel Attention for DNNs in Intelligent Edge Computing (2011.03891v2)

Published 8 Nov 2020 in cs.CV and cs.AI

Abstract: Deep Neural Networks (DNNs) have achieved remarkable success in many computer vision tasks recently, but the huge number of parameters and the high computation overhead hinder their deployments on resource-constrained edge devices. It is worth noting that channel pruning is an effective approach for compressing DNN models. A critical challenge is to determine which channels are to be removed, so that the model accuracy will not be negatively affected. In this paper, we first propose Spatial and Channel Attention (SCA), a new attention module combining both spatial and channel attention that respectively focuses on "where" and "what" are the most informative parts. Guided by the scale values generated by SCA for measuring channel importance, we further propose a new channel pruning approach called Channel Pruning guided by Spatial and Channel Attention (CPSCA). Experimental results indicate that SCA achieves the best inference accuracy, while incurring negligibly extra resource consumption, compared to other state-of-the-art attention modules. Our evaluation on two benchmark datasets shows that, with the guidance of SCA, our CPSCA approach achieves higher inference accuracy than other state-of-the-art pruning methods under the same pruning ratios.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Mengran Liu (1 paper)
  2. Weiwei Fang (2 papers)
  3. Xiaodong Ma (6 papers)
  4. Wenyuan Xu (35 papers)
  5. Naixue Xiong (16 papers)
  6. Yi Ding (92 papers)
Citations (16)

Summary

We haven't generated a summary for this paper yet.