Structured Network Pruning by Measuring Filter-wise Interactions (2307.00758v1)
Abstract: Structured network pruning is a practical approach to reduce computation cost directly while retaining the CNNs' generalization performance in real applications. However, identifying redundant filters is a core problem in structured network pruning, and current redundancy criteria only focus on individual filters' attributes. When pruning sparsity increases, these redundancy criteria are not effective or efficient enough. Since the filter-wise interaction also contributes to the CNN's prediction accuracy, we integrate the filter-wise interaction into the redundancy criterion. In our criterion, we introduce the filter importance and filter utilization strength to reflect the decision ability of individual and multiple filters. Utilizing this new redundancy criterion, we propose a structured network pruning approach SNPFI (Structured Network Pruning by measuring Filter-wise Interaction). During the pruning, the SNPFI can automatically assign the proper sparsity based on the filter utilization strength and eliminate the useless filters by filter importance. After the pruning, the SNPFI can recover pruned model's performance effectively without iterative training by minimizing the interaction difference. We empirically demonstrate the effectiveness of the SNPFI with several commonly used CNN models, including AlexNet, MobileNetv1, and ResNet-50, on various image classification datasets, including MNIST, CIFAR-10, and ImageNet. For all experimental CNN models, nearly 60% of computation is reduced in a network compression while the classification accuracy remains.
- Combining weight pruning and knowledge distillation for cnn compression. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 3191–3198, 2021.
- Shapley value as principled metric for structured network pruning. arXiv preprint arXiv:2006.01795, 2020.
- Structured pruning of deep convolutional neural networks. ACM Journal on Emerging Technologies in Computing Systems (JETC), 13(3):1–18, 2017.
- Rudder: Return decomposition for delayed rewards. Advances in Neural Information Processing Systems, 32, 2019.
- Ensemble knowledge distillation for learning improved and efficient networks. arXiv preprint arXiv:1909.08097, 2019.
- Learning compact representations of neural networks using discriminative masking (dam). Advances in Neural Information Processing Systems, 34:3491–3503, 2021.
- Cooperative game theory: Basic concepts and computational challenges. IEEE Intelligent Systems, 27(3):86–90, 2012.
- Towards efficient model compression via learned global ranking. arXiv, 2019.
- Learning global pairwise interactions with bayesian neural networks. In European Conference on Artificial Intelligence, 2019.
- Discovering and explaining the representation bottleneck of dnns. arXiv preprint arXiv:2111.06236, 2021.
- Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248–255. Ieee, 2009.
- Xuanyi Dong and Yi Yang. Network Pruning via Transformable Architecture Search. arXiv e-prints, art. arXiv:1905.09717, May 2019. doi: 10.48550/arXiv.1905.09717.
- Effective data generation for imbalanced learning using conditional generative adversarial networks. Expert Systems with Application, 2018.
- Network pruning via performance maximization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9270–9280, 2021.
- An axiomatic approach to the concept of interaction among players in cooperative games. International Journal of Game Theory, 28:547–565, 1999.
- Learning both weights and connections for efficient neural network. Advances in neural information processing systems, 28, 2015.
- Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
- Soft filter pruning for accelerating deep convolutional neural networks. arXiv preprint arXiv:1808.06866, 2018a.
- Filter pruning via geometric median for deep convolutional neural networks acceleration. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 4340–4349, 2019.
- Amc: Automl for model compression and acceleration on mobile devices. In Proceedings of the European conference on computer vision (ECCV), pages 784–800, 2018b.
- Heaton and Jeff. Ian goodfellow, yoshua bengio, and aaron courville: Deep learning. Genetic Programming and Evolvable Machines, pages 1–3, 2017.
- Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531, 2015.
- Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861, 2017.
- S. Ioffe and C. Szegedy. Batch normalization: Accelerating deep network training by reducing internal covariate shift. JMLR.org, 2015.
- Explaining explanations: Axiomatic feature interactions for deep networks. The Journal of Machine Learning Research, 22(1):4687–4740, 2021.
- Llvip: A visible-infrared paired dataset for low-light vision. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 3496–3504, 2021.
- Learning multiple layers of features from tiny images. 2009.
- Imagenet classification with deep convolutional neural networks. Communications of the ACM, 60(6):84–90, 2017.
- Optimal brain damage. Advances in neural information processing systems, 2, 1989.
- Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278–2324, 1998.
- Layer-adaptive sparsity for the magnitude-based pruning. arXiv preprint arXiv:2010.07611, 2020.
- Continuous control with deep reinforcement learning. arXiv preprint arXiv:1509.02971, 2015.
- Channel pruning via automatic structure search. arXiv preprint arXiv:2001.08565, 2020.
- Learning efficient convolutional networks through network slimming. In Proceedings of the IEEE international conference on computer vision, pages 2736–2744, 2017.
- S. Lundberg and S. I. Lee. A unified approach to interpreting model predictions. In Nips, 2017.
- Quantisation and pruning for neural network compression and regularisation. In 2020 International SAUPEC/RobMech/PRASA Conference, pages 1–6, 2020. doi: 10.1109/SAUPEC/RobMech/PRASA48453.2020.9041096.
- The shapley value in machine learning. arXiv preprint arXiv:2202.05594, 2022.
- Sebastian Ruder. An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747, 2016.
- Claude E. Shannon. A mathematical theory of communication. Bell Syst. Tech. J., 27:623–656, 1948.
- Hierarchical interpretations for neural network predictions. arXiv preprint arXiv:1806.05337, 2018.
- Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. In ICLR, 2016.
- Axiomatic attribution for deep networks. In International conference on machine learning, pages 3319–3328. PMLR, 2017.
- On the importance of initialization and momentum in deep learning. In International conference on machine learning, pages 1139–1147. PMLR, 2013.
- Detecting statistical interactions from neural network weights. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings. OpenReview.net, 2018. URL https://openreview.net/forum?id=ByOfBggRZ.
- Neural pruning via growing regularization. In International Conference on Learning Representations, 2021.
- Taking human out of learning applications: A survey on automated machine learning. arXiv preprint arXiv:1810.13306, 2018.
- Rethinking the smaller-norm-less-informative assumption in channel pruning of convolution layers. In 6th International Conference on Learning Representations, ICLR 2018,Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings. OpenReview.net, 2018.
- Deep neural networks on chip-a survey. In 2020 IEEE International Conference on Big Data and Smart Computing (BigComp), pages 589–592. IEEE, 2020.
- Topology-aware network pruning using multi-stage graph embedding and reinforcement learning. In International Conference on Machine Learning, pages 25656–25667. PMLR, 2022.
- Image super-resolution using very deep residual channel attention networks. In Proceedings of the European conference on computer vision (ECCV), pages 286–301, 2018.
- Revisit kernel pruning with lottery regulated grouped convolutions. In International Conference on Learning Representations, 2022.
- Neural network distiller: A python package for dnn compression research. arXiv preprint arXiv:1910.12232, 2019.
- Convolutional neural network simplification via feature map pruning. Comput. Electr. Eng., 70:950–958, 2018.
- Wenting Tang (2 papers)
- Xingxing Wei (60 papers)
- Bo Li (1107 papers)