Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
126 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

High-Performance Temporal Reversible Spiking Neural Networks with $O(L)$ Training Memory and $O(1)$ Inference Cost (2405.16466v1)

Published 26 May 2024 in cs.NE

Abstract: Multi-timestep simulation of brain-inspired Spiking Neural Networks (SNNs) boost memory requirements during training and increase inference energy cost. Current training methods cannot simultaneously solve both training and inference dilemmas. This work proposes a novel Temporal Reversible architecture for SNNs (T-RevSNN) to jointly address the training and inference challenges by altering the forward propagation of SNNs. We turn off the temporal dynamics of most spiking neurons and design multi-level temporal reversible interactions at temporal turn-on spiking neurons, resulting in a $O(L)$ training memory. Combined with the temporal reversible nature, we redesign the input encoding and network organization of SNNs to achieve $O(1)$ inference energy cost. Then, we finely adjust the internal units and residual connections of the basic SNN block to ensure the effectiveness of sparse temporal information interaction. T-RevSNN achieves excellent accuracy on ImageNet, while the memory efficiency, training time acceleration, and inference energy efficiency can be significantly improved by $8.6 \times$, $2.0 \times$, and $1.6 \times$, respectively. This work is expected to break the technical bottleneck of significantly increasing memory cost and training time for large-scale SNNs while maintaining high performance and low inference energy cost. Source code and models are available at: https://github.com/BICLab/T-RevSNN.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (85)
  1. A low power, fully event-based gesture recognition system. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.  7243–7252, 2017.
  2. Rezero is all you need: Fast convergence at large depth. In Uncertainty in Artificial Intelligence, pp.  1352–1361. PMLR, 2021.
  3. High-performance large-scale image recognition without normalization. In International Conference on Machine Learning, pp.  1059–1071. PMLR, 2021.
  4. A partially reversible u-net for memory-efficient volumetric image segmentation. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2019: 22nd International Conference, Shenzhen, China, October 13–17, 2019, Proceedings, Part III 22, pp.  429–437. Springer, 2019.
  5. Reversible column networks. In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=Oc2vlWU0jFY.
  6. One timestep is all you need: Training spiking neural networks with ultra low latency. arXiv preprint arXiv:2110.05929, 2021.
  7. Loihi: A neuromorphic manycore processor with on-chip learning. IEEE Micro, 38(1):82–99, 2018.
  8. Tensor decomposition based attention module for spiking neural networks. Knowledge-Based Systems, pp.  111780, 2024.
  9. Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.  248–255, 2009.
  10. Rethinking the performance comparison between snns and anns. Neural Networks, 121:294–307, 2020.
  11. Temporal efficient training of spiking neural network via gradient re-weighting. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=_XNtisL32jv.
  12. Shrinking your timestep: Towards low-latency neuromorphic object recognition with spiking neural network. In Proceedings of the AAAI Conference on Artificial Intelligence, 2024.
  13. An image is worth 16x16 words: Transformers for image recognition at scale. In International Conference on Learning Representations, 2021.
  14. Temporal effective batch normalization in spiking neural networks. Advances in Neural Information Processing Systems, 35:34377–34390, 2022.
  15. Training spiking neural networks using lessons from deep learning. Proceedings of the IEEE, 111(9):1016–1054, 2023.
  16. Deep residual learning in spiking neural networks. Advances in Neural Information Processing Systems, 34:21056–21069, 2021a.
  17. Incorporating learnable membrane time constant to enhance learning of spiking neural networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pp.  2661–2671, October 2021b.
  18. Spikingjelly: An open-source machine learning infrastructure platform for spike-based intelligence. Science Advances, 9(40):eadi1480, 2023.
  19. Event-based vision: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(1):154–180, 2022.
  20. Res2net: A new multi-scale backbone architecture. IEEE transactions on pattern analysis and machine intelligence, 43(2):652–662, 2019.
  21. The reversible residual network: Backpropagation without storing activations. Advances in Neural Information Processing Systems, 30, 2017.
  22. IM-loss: Information maximization loss for spiking neural networks. In Oh, A. H., Agarwal, A., Belgrave, D., and Cho, K. (eds.), Advances in Neural Information Processing Systems, 2022a. URL https://openreview.net/forum?id=Jw34v_84m2b.
  23. Reducing information loss for spiking neural networks. In European Conference on Computer Vision, pp.  36–52. Springer, 2022b.
  24. Recdis-snn: Rectifying membrane potential distribution for directly training spiking neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  326–335, 2022c.
  25. Direct learning-based deep spiking neural networks: a review. Frontiers in Neuroscience, 17:1209795, 2023.
  26. Identity mappings in deep residual networks. In Leibe, B., Matas, J., Sebe, N., and Welling, M. (eds.), Computer Vision – ECCV 2016, pp.  630–645, Cham, 2016a. Springer International Publishing.
  27. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp.  770–778, 2016b.
  28. Horowitz, M. 1.1 computing’s energy problem (and what we can do about it). In 2014 IEEE International Solid-State Circuits Conference Digest of Technical Papers (ISSCC), pp.  10–14. IEEE, 2014.
  29. Spiking deep residual networks. IEEE Transactions on Neural Networks and Learning Systems, 34(8):5200–5205, 2023a. doi: 10.1109/TNNLS.2021.3119238.
  30. Fast-snn: Fast spiking neural network by converting quantized ann. arXiv preprint arXiv:2305.19868, 2023b.
  31. Advancing spiking neural networks toward deep residual learning. IEEE Transactions on Neural Networks and Learning Systems, 2024.
  32. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In International Conference on Machine Learning, pp.  448–456. PMLR, 2015.
  33. Revisiting batch normalization for training low-latency deep spiking neural networks from scratch. Frontiers in Neuroscience, 15:773954, 2021.
  34. Rate coding or direct coding: Which one is better for accurate, robust, and energy-efficient spiking neural networks? In ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp.  71–75, 2022.
  35. Efficient processing of spatio-temporal data streams with spiking neural networks. Frontiers in Neuroscience, 14:439, 2020.
  36. Cifar10-dvs: an event-stream dataset for object classification. Frontiers in Neuroscience, 11:309, 2017.
  37. Differentiable spike: Rethinking gradient-descent for training spiking neural networks. Advances in Neural Information Processing Systems, 34:23426–23439, 2021.
  38. Seenn: Towards temporal spiking early-exit neural networks. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
  39. Learnable surrogate gradient for direct training spiking neural networks. In Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, IJCAI-23, pp.  3002–3010, 8 2023.
  40. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July 2017.
  41. Rethinking pretraining as a bridge from anns to snns. IEEE Transactions on Neural Networks and Learning Systems, pp.  1–14, 2022. doi: 10.1109/TNNLS.2022.3217796.
  42. A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  11976–11986, 2022.
  43. Maass, W. Networks of spiking neurons: the third generation of neural network models. Neural Networks, 10(9):1659–1671, 1997.
  44. Reversible recurrent neural networks. Advances in Neural Information Processing Systems, 31, 2018.
  45. Reversible vision transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  10830–10840, 2022.
  46. Training high-performance low-latency spiking neural networks by differentiation on spike representation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  12444–12453, 2022.
  47. Towards memory-and time-efficient backpropagation for training spiking neural networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp.  6166–6176, 2023.
  48. A million spiking-neuron integrated circuit with a scalable communication network and interface. Science, 345(6197):668–673, 2014.
  49. Surrogate gradient learning in spiking neural networks: Bringing the power of gradient-based optimization to spiking neural networks. IEEE Signal Processing Magazine, 36(6):51–63, 2019.
  50. Toward scalable, efficient, and accurate deep spiking neural networks with backward residual connections, stochastic softmax, and hybridization. Frontiers in Neuroscience, 14:653, 2020.
  51. Towards artificial general intelligence with hybrid tianjic chip architecture. Nature, 572(7767):106–111, 2019.
  52. Sparse spiking gradient descent. Advances in Neural Information Processing Systems, 34:11795–11808, 2021.
  53. Gated attention coding for training high-performance and efficient spiking neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 38, pp.  601–610, 2024.
  54. Vtsnn: A virtual temporal spiking neural network. Frontiers in neuroscience, 17:1091097, 2023.
  55. Diet-snn: A low-latency spiking neural network with direct input encoding and leakage and threshold optimization. IEEE Transactions on Neural Networks and Learning Systems, 34(6):3174–3182, 2023. doi: 10.1109/TNNLS.2021.3111897.
  56. Enabling deep spiking neural networks with hybrid conversion and spike timing dependent backpropagation. In International Conference on Learning Representations, 2020. URL https://openreview.net/forum?id=B1xSperKvH.
  57. Towards spike-based machine intelligence with neuromorphic computing. Nature, 575(7784):607–617, 2019.
  58. Momentum residual neural networks. In International Conference on Machine Learning, pp.  9276–9287. PMLR, 2021.
  59. Opportunities for neuromorphic computing algorithms and applications. Nature Computational Science, 2(1):10–19, 2022.
  60. Going deeper in spiking neural networks: Vgg and residual architectures. Frontiers in Neuroscience, 13:95, 2019.
  61. Sequence approximation using feedforward spiking neural network for spatiotemporal learning: Theory and optimization methods. In International Conference on Learning Representations, 2021.
  62. Attention is all you need. In Advances in Neural Information Processing Systems, pp.  5998–6008, 2017.
  63. Kronecker cp decomposition with fast multiplication for compressing rnns. IEEE Transactions on Neural Networks and Learning Systems, 34(5):2205–2219, 2023a.
  64. Masked spiking transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp.  1761–1771, 2023b.
  65. Wightman, R. Pytorch image models. https://github.com/rwightman/pytorch-image-models, 2019.
  66. Progressive tandem learning for pattern recognition with deep spiking neural networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(11):7824–7840, 2021a.
  67. Mss-depthnet: Depth prediction with multi-step spiking neural network. arXiv preprint arXiv:2211.12156, 2022a.
  68. Spatio-temporal backpropagation for training high-performance spiking neural networks. Frontiers in Neuroscience, 12:331, 2018.
  69. Direct training for spiking neural networks: Faster, larger, better. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pp.  1311–1318, 2019.
  70. Efficient visual recognition: A survey on recent advances and brain-inspired methodologies. Machine Intelligence Research, 19(5):366–411, 2022b.
  71. Liaf-net: Leaky integrate and analog fire network for lightweight and efficient spatiotemporal information processing. IEEE Transactions on Neural Networks and Learning Systems, pp.  1–14, 2021b. doi: 10.1109/TNNLS.2021.3073016.
  72. Online training through time for spiking neural networks. Advances in Neural Information Processing Systems, 35:20717–20730, 2022.
  73. Transformers in computational visual media: A survey. Computational Visual Media, 8:33–62, 2022.
  74. Lead federated neuromorphic learning for wireless edge artificial intelligence. Nature Communications, 13(1):4269, 2022.
  75. Temporal-wise attention spiking neural networks for event streams classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp.  10221–10230, 2021.
  76. Inherent redundancy in spiking neural networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp.  16924–16934, 2023a.
  77. Spike-driven transformer. In Thirty-seventh Conference on Neural Information Processing Systems, 2023b. URL https://openreview.net/forum?id=9FmolyOHi5.
  78. Sparser spiking activity can be better: Feature refine-and-mask spiking neural network for event-based visual recognition. Neural Networks, 166:410–423, 2023c.
  79. Attention spiking neural networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(8):9393–9410, 2023d.
  80. Spike-driven transformer v2: Meta spiking neural network architecture inspiring the design of next-generation neuromorphic chips. In The Twelfth International Conference on Learning Representations, 2024a. URL https://openreview.net/forum?id=1SIBN5Xyw7.
  81. Spike-based dynamic computing with asynchronous sensing-computing neuromorphic chip. Nature Communications, 15(1):4464, May 2024b. ISSN 2041-1723. URL https://doi.org/10.1038/s41467-024-47811-6.
  82. Accurate and efficient time-domain classification with adaptive spiking recurrent neural networks. Nature Machine Intelligence, 3(10):905–913, 2021.
  83. Memory-efficient reversible spiking neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, 2024.
  84. Going deeper with directly-trained larger spiking neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pp.  11062–11070, 2021.
  85. Spikformer: When spiking neural network meets transformer. In The Eleventh International Conference on Learning Representations, 2023.
Citations (4)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets