High-Performance Temporal Reversible Spiking Neural Networks with $O(L)$ Training Memory and $O(1)$ Inference Cost (2405.16466v1)
Abstract: Multi-timestep simulation of brain-inspired Spiking Neural Networks (SNNs) boost memory requirements during training and increase inference energy cost. Current training methods cannot simultaneously solve both training and inference dilemmas. This work proposes a novel Temporal Reversible architecture for SNNs (T-RevSNN) to jointly address the training and inference challenges by altering the forward propagation of SNNs. We turn off the temporal dynamics of most spiking neurons and design multi-level temporal reversible interactions at temporal turn-on spiking neurons, resulting in a $O(L)$ training memory. Combined with the temporal reversible nature, we redesign the input encoding and network organization of SNNs to achieve $O(1)$ inference energy cost. Then, we finely adjust the internal units and residual connections of the basic SNN block to ensure the effectiveness of sparse temporal information interaction. T-RevSNN achieves excellent accuracy on ImageNet, while the memory efficiency, training time acceleration, and inference energy efficiency can be significantly improved by $8.6 \times$, $2.0 \times$, and $1.6 \times$, respectively. This work is expected to break the technical bottleneck of significantly increasing memory cost and training time for large-scale SNNs while maintaining high performance and low inference energy cost. Source code and models are available at: https://github.com/BICLab/T-RevSNN.
- A low power, fully event-based gesture recognition system. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7243–7252, 2017.
- Rezero is all you need: Fast convergence at large depth. In Uncertainty in Artificial Intelligence, pp. 1352–1361. PMLR, 2021.
- High-performance large-scale image recognition without normalization. In International Conference on Machine Learning, pp. 1059–1071. PMLR, 2021.
- A partially reversible u-net for memory-efficient volumetric image segmentation. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2019: 22nd International Conference, Shenzhen, China, October 13–17, 2019, Proceedings, Part III 22, pp. 429–437. Springer, 2019.
- Reversible column networks. In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=Oc2vlWU0jFY.
- One timestep is all you need: Training spiking neural networks with ultra low latency. arXiv preprint arXiv:2110.05929, 2021.
- Loihi: A neuromorphic manycore processor with on-chip learning. IEEE Micro, 38(1):82–99, 2018.
- Tensor decomposition based attention module for spiking neural networks. Knowledge-Based Systems, pp. 111780, 2024.
- Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 248–255, 2009.
- Rethinking the performance comparison between snns and anns. Neural Networks, 121:294–307, 2020.
- Temporal efficient training of spiking neural network via gradient re-weighting. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=_XNtisL32jv.
- Shrinking your timestep: Towards low-latency neuromorphic object recognition with spiking neural network. In Proceedings of the AAAI Conference on Artificial Intelligence, 2024.
- An image is worth 16x16 words: Transformers for image recognition at scale. In International Conference on Learning Representations, 2021.
- Temporal effective batch normalization in spiking neural networks. Advances in Neural Information Processing Systems, 35:34377–34390, 2022.
- Training spiking neural networks using lessons from deep learning. Proceedings of the IEEE, 111(9):1016–1054, 2023.
- Deep residual learning in spiking neural networks. Advances in Neural Information Processing Systems, 34:21056–21069, 2021a.
- Incorporating learnable membrane time constant to enhance learning of spiking neural networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pp. 2661–2671, October 2021b.
- Spikingjelly: An open-source machine learning infrastructure platform for spike-based intelligence. Science Advances, 9(40):eadi1480, 2023.
- Event-based vision: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(1):154–180, 2022.
- Res2net: A new multi-scale backbone architecture. IEEE transactions on pattern analysis and machine intelligence, 43(2):652–662, 2019.
- The reversible residual network: Backpropagation without storing activations. Advances in Neural Information Processing Systems, 30, 2017.
- IM-loss: Information maximization loss for spiking neural networks. In Oh, A. H., Agarwal, A., Belgrave, D., and Cho, K. (eds.), Advances in Neural Information Processing Systems, 2022a. URL https://openreview.net/forum?id=Jw34v_84m2b.
- Reducing information loss for spiking neural networks. In European Conference on Computer Vision, pp. 36–52. Springer, 2022b.
- Recdis-snn: Rectifying membrane potential distribution for directly training spiking neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 326–335, 2022c.
- Direct learning-based deep spiking neural networks: a review. Frontiers in Neuroscience, 17:1209795, 2023.
- Identity mappings in deep residual networks. In Leibe, B., Matas, J., Sebe, N., and Welling, M. (eds.), Computer Vision – ECCV 2016, pp. 630–645, Cham, 2016a. Springer International Publishing.
- Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778, 2016b.
- Horowitz, M. 1.1 computing’s energy problem (and what we can do about it). In 2014 IEEE International Solid-State Circuits Conference Digest of Technical Papers (ISSCC), pp. 10–14. IEEE, 2014.
- Spiking deep residual networks. IEEE Transactions on Neural Networks and Learning Systems, 34(8):5200–5205, 2023a. doi: 10.1109/TNNLS.2021.3119238.
- Fast-snn: Fast spiking neural network by converting quantized ann. arXiv preprint arXiv:2305.19868, 2023b.
- Advancing spiking neural networks toward deep residual learning. IEEE Transactions on Neural Networks and Learning Systems, 2024.
- Batch normalization: Accelerating deep network training by reducing internal covariate shift. In International Conference on Machine Learning, pp. 448–456. PMLR, 2015.
- Revisiting batch normalization for training low-latency deep spiking neural networks from scratch. Frontiers in Neuroscience, 15:773954, 2021.
- Rate coding or direct coding: Which one is better for accurate, robust, and energy-efficient spiking neural networks? In ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 71–75, 2022.
- Efficient processing of spatio-temporal data streams with spiking neural networks. Frontiers in Neuroscience, 14:439, 2020.
- Cifar10-dvs: an event-stream dataset for object classification. Frontiers in Neuroscience, 11:309, 2017.
- Differentiable spike: Rethinking gradient-descent for training spiking neural networks. Advances in Neural Information Processing Systems, 34:23426–23439, 2021.
- Seenn: Towards temporal spiking early-exit neural networks. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
- Learnable surrogate gradient for direct training spiking neural networks. In Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, IJCAI-23, pp. 3002–3010, 8 2023.
- Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July 2017.
- Rethinking pretraining as a bridge from anns to snns. IEEE Transactions on Neural Networks and Learning Systems, pp. 1–14, 2022. doi: 10.1109/TNNLS.2022.3217796.
- A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11976–11986, 2022.
- Maass, W. Networks of spiking neurons: the third generation of neural network models. Neural Networks, 10(9):1659–1671, 1997.
- Reversible recurrent neural networks. Advances in Neural Information Processing Systems, 31, 2018.
- Reversible vision transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10830–10840, 2022.
- Training high-performance low-latency spiking neural networks by differentiation on spike representation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 12444–12453, 2022.
- Towards memory-and time-efficient backpropagation for training spiking neural networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 6166–6176, 2023.
- A million spiking-neuron integrated circuit with a scalable communication network and interface. Science, 345(6197):668–673, 2014.
- Surrogate gradient learning in spiking neural networks: Bringing the power of gradient-based optimization to spiking neural networks. IEEE Signal Processing Magazine, 36(6):51–63, 2019.
- Toward scalable, efficient, and accurate deep spiking neural networks with backward residual connections, stochastic softmax, and hybridization. Frontiers in Neuroscience, 14:653, 2020.
- Towards artificial general intelligence with hybrid tianjic chip architecture. Nature, 572(7767):106–111, 2019.
- Sparse spiking gradient descent. Advances in Neural Information Processing Systems, 34:11795–11808, 2021.
- Gated attention coding for training high-performance and efficient spiking neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 38, pp. 601–610, 2024.
- Vtsnn: A virtual temporal spiking neural network. Frontiers in neuroscience, 17:1091097, 2023.
- Diet-snn: A low-latency spiking neural network with direct input encoding and leakage and threshold optimization. IEEE Transactions on Neural Networks and Learning Systems, 34(6):3174–3182, 2023. doi: 10.1109/TNNLS.2021.3111897.
- Enabling deep spiking neural networks with hybrid conversion and spike timing dependent backpropagation. In International Conference on Learning Representations, 2020. URL https://openreview.net/forum?id=B1xSperKvH.
- Towards spike-based machine intelligence with neuromorphic computing. Nature, 575(7784):607–617, 2019.
- Momentum residual neural networks. In International Conference on Machine Learning, pp. 9276–9287. PMLR, 2021.
- Opportunities for neuromorphic computing algorithms and applications. Nature Computational Science, 2(1):10–19, 2022.
- Going deeper in spiking neural networks: Vgg and residual architectures. Frontiers in Neuroscience, 13:95, 2019.
- Sequence approximation using feedforward spiking neural network for spatiotemporal learning: Theory and optimization methods. In International Conference on Learning Representations, 2021.
- Attention is all you need. In Advances in Neural Information Processing Systems, pp. 5998–6008, 2017.
- Kronecker cp decomposition with fast multiplication for compressing rnns. IEEE Transactions on Neural Networks and Learning Systems, 34(5):2205–2219, 2023a.
- Masked spiking transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 1761–1771, 2023b.
- Wightman, R. Pytorch image models. https://github.com/rwightman/pytorch-image-models, 2019.
- Progressive tandem learning for pattern recognition with deep spiking neural networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(11):7824–7840, 2021a.
- Mss-depthnet: Depth prediction with multi-step spiking neural network. arXiv preprint arXiv:2211.12156, 2022a.
- Spatio-temporal backpropagation for training high-performance spiking neural networks. Frontiers in Neuroscience, 12:331, 2018.
- Direct training for spiking neural networks: Faster, larger, better. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pp. 1311–1318, 2019.
- Efficient visual recognition: A survey on recent advances and brain-inspired methodologies. Machine Intelligence Research, 19(5):366–411, 2022b.
- Liaf-net: Leaky integrate and analog fire network for lightweight and efficient spatiotemporal information processing. IEEE Transactions on Neural Networks and Learning Systems, pp. 1–14, 2021b. doi: 10.1109/TNNLS.2021.3073016.
- Online training through time for spiking neural networks. Advances in Neural Information Processing Systems, 35:20717–20730, 2022.
- Transformers in computational visual media: A survey. Computational Visual Media, 8:33–62, 2022.
- Lead federated neuromorphic learning for wireless edge artificial intelligence. Nature Communications, 13(1):4269, 2022.
- Temporal-wise attention spiking neural networks for event streams classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 10221–10230, 2021.
- Inherent redundancy in spiking neural networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 16924–16934, 2023a.
- Spike-driven transformer. In Thirty-seventh Conference on Neural Information Processing Systems, 2023b. URL https://openreview.net/forum?id=9FmolyOHi5.
- Sparser spiking activity can be better: Feature refine-and-mask spiking neural network for event-based visual recognition. Neural Networks, 166:410–423, 2023c.
- Attention spiking neural networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(8):9393–9410, 2023d.
- Spike-driven transformer v2: Meta spiking neural network architecture inspiring the design of next-generation neuromorphic chips. In The Twelfth International Conference on Learning Representations, 2024a. URL https://openreview.net/forum?id=1SIBN5Xyw7.
- Spike-based dynamic computing with asynchronous sensing-computing neuromorphic chip. Nature Communications, 15(1):4464, May 2024b. ISSN 2041-1723. URL https://doi.org/10.1038/s41467-024-47811-6.
- Accurate and efficient time-domain classification with adaptive spiking recurrent neural networks. Nature Machine Intelligence, 3(10):905–913, 2021.
- Memory-efficient reversible spiking neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, 2024.
- Going deeper with directly-trained larger spiking neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pp. 11062–11070, 2021.
- Spikformer: When spiking neural network meets transformer. In The Eleventh International Conference on Learning Representations, 2023.