Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Boosting Pruned Networks with Linear Over-parameterization (2204.11444v3)

Published 25 Apr 2022 in cs.CV

Abstract: Structured pruning compresses neural networks by reducing channels (filters) for fast inference and low footprint at run-time. To restore accuracy after pruning, fine-tuning is usually applied to pruned networks. However, too few remaining parameters in pruned networks inevitably bring a great challenge to fine-tuning to restore accuracy. To address this challenge, we propose a novel method that first linearly over-parameterizes the compact layers in pruned networks to enlarge the number of fine-tuning parameters and then re-parameterizes them to the original layers after fine-tuning. Specifically, we equivalently expand the convolution/linear layer with several consecutive convolution/linear layers that do not alter the current output feature maps. Furthermore, we utilize similarity-preserving knowledge distillation that encourages the over-parameterized block to learn the immediate data-to-data similarities of the corresponding dense layer to maintain its feature learning ability. The proposed method is comprehensively evaluated on CIFAR-10 and ImageNet which significantly outperforms the vanilla fine-tuning strategy, especially for large pruning ratio.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (21)
  1. “Network pruning via performance maximization,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 9270–9280.
  2. “Pruning and quantization for deep neural network acceleration: A survey,” Neurocomputing, vol. 461, pp. 370–403, 2021.
  3. “Accelerate cnns from three dimensions: A comprehensive pruning framework,” in International Conference on Machine Learning. PMLR, 2021, pp. 10717–10726.
  4. “Nisp: Pruning networks using neuron importance score propagation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 9194–9203.
  5. “Centripetal sgd for pruning very deep convolutional networks with complicated structure,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 4943–4953.
  6. “Pruning filters for efficient convnets,” arXiv preprint arXiv:1608.08710, 2016.
  7. “Learning to prune deep neural networks via layer-wise optimal brain surgeon,” Advances in Neural Information Processing Systems, vol. 30, 2017.
  8. “A signal propagation perspective for pruning neural networks at initialization,” arXiv preprint arXiv:1906.06307, 2019.
  9. “Channel pruning for accelerating very deep neural networks,” in Proceedings of the IEEE international conference on computer vision, 2017.
  10. “To prune, or not to prune: exploring the efficacy of pruning for model compression,” arXiv preprint arXiv:1710.01878, 2017.
  11. “Expandnets: Linear over-parameterization to train compact convolutional networks,” arXiv preprint arXiv:1811.10495, 2018.
  12. “Repvgg: Making vgg-style convnets great again,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021.
  13. “Similarity-preserving knowledge distillation,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019.
  14. “DENSE: Data-free one-shot federated learning,” in Advances in Neural Information Processing Systems, Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, Eds., 2022.
  15. “The rank of a random matrix,” Applied mathematics and computation, vol. 185, no. 1, pp. 689–694, 2007.
  16. “Learning multiple layers of features from tiny images,” Tech. Rep., 2009.
  17. “Imagenet: A large-scale hierarchical image database,” in 2009 IEEE conference on computer vision and pattern recognition. Ieee, 2009.
  18. “Amc: Automl for model compression and acceleration on mobile devices,” in Proceedings of the European conference on computer vision (ECCV), 2018, pp. 784–800.
  19. “Hrank: Filter pruning using high-rank feature map,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 1529–1538.
  20. “Thinet: A filter level pruning method for deep neural network compression,” in Proceedings of the IEEE international conference on computer vision, 2017.
  21. “Eagleeye: Fast sub-net evaluation for efficient neural network pruning,” in European conference on computer vision. Springer, 2020, pp. 639–654.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Yu Qian (18 papers)
  2. Jian Cao (81 papers)
  3. Xiaoshuang Li (5 papers)
  4. Jie Zhang (847 papers)
  5. Hufei Li (1 paper)
  6. Jue Chen (5 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.