Brain-Inspired Efficient Pruning: Exploiting Criticality in Spiking Neural Networks (2311.16141v3)
Abstract: Spiking Neural Networks (SNNs) have gained significant attention due to the energy-efficient and multiplication-free characteristics. Despite these advantages, deploying large-scale SNNs on edge hardware is challenging due to limited resource availability. Network pruning offers a viable approach to compress the network scale and reduce hardware resource requirements for model deployment. However, existing SNN pruning methods cause high pruning costs and performance loss because they lack efficiency in processing the sparse spike representation of SNNs. In this paper, inspired by the critical brain hypothesis in neuroscience and the high biological plausibility of SNNs, we explore and leverage criticality to facilitate efficient pruning in deep SNNs. We firstly explain criticality in SNNs from the perspective of maximizing feature information entropy. Second, We propose a low-cost metric for assess neuron criticality in feature transmission and design a pruning-regeneration method that incorporates this criticality into the pruning process. Experimental results demonstrate that our method achieves higher performance than the current state-of-the-art (SOTA) method with up to 95.26\% reduction of pruning cost. The criticality-based regeneration process efficiently selects potential structures and facilitates consistent feature representation.
- Truenorth: Design and tool flow of a 65 mw 1 million neuron programmable neurosynaptic chip. IEEE transactions on computer-aided design of integrated circuits and systems, 34(10): 1537–1557.
- Beggs, J. M. 2008. The criticality hypothesis: how local cortical networks might optimize information processing. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences, 366(1864): 329–343.
- Being critical of criticality in the brain. Frontiers in physiology, 3: 163.
- Long short-term memory and learning-to-learn in networks of spiking neurons. Advances in neural information processing systems, 31.
- Pruning of Deep Spiking Neural Networks through Gradient Rewiring. In Zhou, Z.-H., ed., Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI-21, 1713–1721. International Joint Conferences on Artificial Intelligence Organization. Main Track.
- Loihi: A neuromorphic manycore processor with on-chip learning. Ieee Micro, 38(1): 82–99.
- Comprehensive snn compression using admm optimization and activity regularization. IEEE transactions on neural networks and learning systems.
- Landau–Ginzburg theory of cortex dynamics: Scale-free avalanches emerge at the edge of synchronization. Proceedings of the National Academy of Sciences, 115(7): E1356–E1365.
- Resrep: Lossless cnn pruning via decoupling remembering and forgetting. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 4510–4520.
- Global sparse momentum sgd for pruning very deep neural networks. Advances in Neural Information Processing Systems, 32.
- SpikingJelly. https://github.com/fangwei123456/spikingjelly. Accessed: 2022-5-1.
- The lottery ticket hypothesis: Finding sparse, trainable neural networks. arXiv preprint arXiv:1803.03635.
- Self-organized criticality in single-neuron excitability. Physical Review E, 88(6): 062717.
- The state of sparsity in deep neural networks. arXiv preprint arXiv:1902.09574.
- Learning both weights and connections for efficient neural network. Advances in neural information processing systems, 28.
- Hansen, L. 2015. Tiny ImageNet challenge submission. CS 231N, 5.
- Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, 770–778.
- Filter pruning via geometric median for deep convolutional neural networks acceleration. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 4340–4349.
- Channel pruning for accelerating very deep neural networks. In Proceedings of the IEEE international conference on computer vision, 1389–1397.
- Criticality, connectivity, and neural disorder: a multifaceted approach to neural computation. Frontiers in computational neuroscience, 15: 611183.
- Earthquake cycles and neural reverberations: collective oscillations in systems with pulse-coupled threshold elements. Physical review letters, 75(6): 1222.
- Self-organized criticality as a fundamental property of neural systems. Frontiers in systems neuroscience, 8: 166.
- Network trimming: A data-driven neuron pruning approach towards efficient deep architectures. arXiv preprint arXiv:1607.03250.
- Izhikevich, E. M. 2003. Simple model of spiking neurons. IEEE Transactions on neural networks, 14(6): 1569–1572.
- Network plasticity as Bayesian inference. PLoS computational biology, 11(11): e1004485.
- Kiang, M. Y. 2001. Extending the Kohonen self-organizing map networks for clustering analysis. Computational Statistics & Data Analysis, 38(2): 161–180.
- Exploring temporal information dynamics in spiking neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, 8308–8316.
- Exploring lottery ticket hypothesis in spiking neural networks. In Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022, Proceedings, Part XII, 102–120. Springer.
- Optimal dynamical range of excitable networks at criticality. Nature physics, 2(5): 348–351.
- Learning multiple layers of features from tiny images.
- Spike-thrift: Towards energy-efficient deep spiking neural networks by limiting spiking activity via attention-guided compression. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 3953–3962.
- Hrank: Filter pruning using high-rank feature map. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 1529–1538.
- Dynsnn: A dynamic approach to reduce redundancy in spiking neural networks. In ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2130–2134. IEEE.
- Sparse training via boosting pruning plasticity with neuroregeneration. Advances in Neural Information Processing Systems, 34: 9908–9922.
- Application of deep compression technique in spiking neural network chip. IEEE transactions on biomedical circuits and systems, 14(2): 274–282.
- Learning Efficient Convolutional Networks Through Network Slimming. In Proceedings of the IEEE International Conference on Computer Vision (ICCV).
- Maass, W. 1997. Networks of spiking neurons: the third generation of neural network models. Neural networks, 10(9): 1659–1671.
- Maass, W. 2014. Noise as a resource for computation and learning in networks of spiking neurons. Proceedings of the IEEE, 102(5): 860–880.
- Importance estimation for neural network pruning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 11264–11272.
- Pruning convolutional neural networks for resource efficient inference. arXiv preprint arXiv:1611.06440.
- Stochastic synapses enable efficient brain-inspired learning machines. Frontiers in neuroscience, 10: 241.
- New K-means clustering methods that minimize the total intra-cluster variance. Afr. J. Math. Stat. Stud, 3: 42–54.
- Towards artificial general intelligence with hybrid Tianjic chip architecture. Nature, 572(7767): 106–111.
- Noise in integrate-and-fire neurons: from stochastic input to escape rates. Neural computation, 12(2): 367–384.
- Jointly learning network connections and link weights in spiking neural networks. In IJCAI, 1597–1603.
- Towards spike-based machine intelligence with neuromorphic computing. Nature, 575(7784): 607–617.
- Neuronal avalanches imply maximum dynamic range in cortical networks at criticality. Journal of neuroscience, 29(49): 15595–15600.
- In-hardware learning of multilayer spiking neural networks on a neuromorphic processor. In 2021 58th ACM/IEEE Design Automation Conference (DAC), 367–372. IEEE.
- Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556.
- Turing, A. M. 2009. Computing machinery and intelligence. Springer.
- Neural Pruning via Growing Regularization. In International Conference on Learning Representations (ICLR).
- Spatio-temporal backpropagation for training high-performance spiking neural networks. Frontiers in neuroscience, 12: 331.
- Direct training for spiking neural networks: Faster, larger, better. In Proceedings of the AAAI conference on artificial intelligence, volume 33, 1311–1318.
- Drawing early-bird tickets: Towards more efficient training of deep networks. arXiv preprint arXiv:1909.11957.
- Going deeper with directly-trained larger spiking neural networks. In Proceedings of the AAAI conference on artificial intelligence, volume 35, 11062–11070.
- To prune, or not to prune: exploring the efficacy of pruning for model compression. arXiv preprint arXiv:1710.01878.