Boosting Pruned Networks with Linear Over-parameterization (2204.11444v3)
Abstract: Structured pruning compresses neural networks by reducing channels (filters) for fast inference and low footprint at run-time. To restore accuracy after pruning, fine-tuning is usually applied to pruned networks. However, too few remaining parameters in pruned networks inevitably bring a great challenge to fine-tuning to restore accuracy. To address this challenge, we propose a novel method that first linearly over-parameterizes the compact layers in pruned networks to enlarge the number of fine-tuning parameters and then re-parameterizes them to the original layers after fine-tuning. Specifically, we equivalently expand the convolution/linear layer with several consecutive convolution/linear layers that do not alter the current output feature maps. Furthermore, we utilize similarity-preserving knowledge distillation that encourages the over-parameterized block to learn the immediate data-to-data similarities of the corresponding dense layer to maintain its feature learning ability. The proposed method is comprehensively evaluated on CIFAR-10 and ImageNet which significantly outperforms the vanilla fine-tuning strategy, especially for large pruning ratio.
- “Network pruning via performance maximization,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 9270–9280.
- “Pruning and quantization for deep neural network acceleration: A survey,” Neurocomputing, vol. 461, pp. 370–403, 2021.
- “Accelerate cnns from three dimensions: A comprehensive pruning framework,” in International Conference on Machine Learning. PMLR, 2021, pp. 10717–10726.
- “Nisp: Pruning networks using neuron importance score propagation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 9194–9203.
- “Centripetal sgd for pruning very deep convolutional networks with complicated structure,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 4943–4953.
- “Pruning filters for efficient convnets,” arXiv preprint arXiv:1608.08710, 2016.
- “Learning to prune deep neural networks via layer-wise optimal brain surgeon,” Advances in Neural Information Processing Systems, vol. 30, 2017.
- “A signal propagation perspective for pruning neural networks at initialization,” arXiv preprint arXiv:1906.06307, 2019.
- “Channel pruning for accelerating very deep neural networks,” in Proceedings of the IEEE international conference on computer vision, 2017.
- “To prune, or not to prune: exploring the efficacy of pruning for model compression,” arXiv preprint arXiv:1710.01878, 2017.
- “Expandnets: Linear over-parameterization to train compact convolutional networks,” arXiv preprint arXiv:1811.10495, 2018.
- “Repvgg: Making vgg-style convnets great again,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021.
- “Similarity-preserving knowledge distillation,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019.
- “DENSE: Data-free one-shot federated learning,” in Advances in Neural Information Processing Systems, Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, Eds., 2022.
- “The rank of a random matrix,” Applied mathematics and computation, vol. 185, no. 1, pp. 689–694, 2007.
- “Learning multiple layers of features from tiny images,” Tech. Rep., 2009.
- “Imagenet: A large-scale hierarchical image database,” in 2009 IEEE conference on computer vision and pattern recognition. Ieee, 2009.
- “Amc: Automl for model compression and acceleration on mobile devices,” in Proceedings of the European conference on computer vision (ECCV), 2018, pp. 784–800.
- “Hrank: Filter pruning using high-rank feature map,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 1529–1538.
- “Thinet: A filter level pruning method for deep neural network compression,” in Proceedings of the IEEE international conference on computer vision, 2017.
- “Eagleeye: Fast sub-net evaluation for efficient neural network pruning,” in European conference on computer vision. Springer, 2020, pp. 639–654.