Papers
Topics
Authors
Recent
2000 character limit reached

Stable Learning Using Spiking Neural Networks Equipped With Affine Encoders and Decoders (2404.04549v3)

Published 6 Apr 2024 in cs.NE, cs.LG, math.FA, and stat.ML

Abstract: We study the learning problem associated with spiking neural networks. Specifically, we focus on spiking neural networks composed of simple spiking neurons having only positive synaptic weights, equipped with an affine encoder and decoder; we refer to these as affine spiking neural networks. These neural networks are shown to depend continuously on their parameters, which facilitates classical covering number-based generalization statements and supports stable gradient-based training. We demonstrate that the positivity of the weights enables a wide range of expressivity results, including rate-optimal approximation of smooth functions and dimension-independent approximation of Barron regular functions. In particular, we show in theory and simulations that affine spiking neural networks are capable of approximating shallow ReLU neural networks. Furthermore, we apply these affine spiking neural networks to standard machine learning benchmarks and reach competitive results. Finally, we observe that from a generalization perspective, contrary to feedforward neural networks or previous results for general spiking neural networks, the depth has little to no adverse effect on the generalization capabilities.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (54)
  1. Moshe Abeles. Role of the cortical neuron: integrator or coincidence detector? Israel journal of medical sciences, 18(1):83–92, 1982.
  2. Neural network learning: Theoretical foundations, volume 9. cambridge university press Cambridge, 1999.
  3. Andrew R Barron. Neural net approximation. In Proc. 7th Yale workshop on adaptive and learning systems, volume 1, pages 69–72, 1992.
  4. Andrew R. Barron. Universal approximation bounds for superpositions of a sigmoidal function. IEEE Trans. Inform. Theory, 39(3):930–945, 1993.
  5. Reconciling modern machine-learning practice and the classical bias–variance trade-off. Proceedings of the National Academy of Sciences, 116(32):15849–15854, 2019.
  6. Deep learning, volume 1. MIT press Cambridge, MA, USA, 2017.
  7. Analysis of the generalization error: Empirical risk minimization over deep artificial neural networks overcomes the curse of dimensionality in the numerical approximation of black–scholes partial differential equations. SIAM Journal on Mathematics of Data Science, 2(3):631–657, 2020.
  8. The modern mathematics of deep learning. arXiv preprint arXiv:2105.04026, pages 86–114, 2021.
  9. Primer of adaptive finite element methods. Multiscale and Adaptivity: Modeling, Numerics and Applications: CIME Summer School, Cetraro, Italy 2009, Editors: Giovanni Naldi, Giovanni Russo, pages 125–225, 2012.
  10. Neural network approximation and estimation of classifiers with classification boundary in a Barron class. The Annals of Applied Probability, 33(4):3039–3079, 2023.
  11. Temporal coding in spiking neural networks with alpha synaptic function. In ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 8529–8533. IEEE, 2020.
  12. Introduction to algorithms. MIT press, 2022.
  13. George Cybenko. Approximation by superpositions of a sigmoidal function. Mathematics of control, signals and systems, 2(4):303–314, 1989.
  14. Advancing mathematics by guiding human intuition with ai. Nature, 600(7887):70–74, 2021.
  15. Representation formulas and pointwise properties for Barron functions. Calculus of Variations and Partial Differential Equations, 61(2):1–37, 2022.
  16. Wulfram Gerstner. Time structure of the activity in neural network models. Physical review E, 51(1):738, 1995.
  17. Neuronal dynamics: From single neurons to networks and models of cognition. Cambridge University Press, 2014.
  18. How to describe neuronal activity: spikes, rates, or assemblies? Advances in neural information processing systems, 6, 1993.
  19. Proof of the theory-to-practice gap in deep learning via sampling complexity bounds for neural network approximation spaces. Foundations of Computational Mathematics, pages 1–59, 2023.
  20. Spiking neural networks: Principles and challenges. In ESANN. Bruges, 2014.
  21. Relu deep neural networks and linear finite elements. Journal of Computational Mathematics, 38(3):502–527, 2020.
  22. A quantitative description of membrane current and its application to conduction and excitation in nerve. The Journal of physiology, 117(4):500, 1952.
  23. A quantitative description of membrane current and its application to conduction and excitation in nerve. Bulletin of mathematical biology, 52:25–71, 1990.
  24. Universal approximation of an unknown mapping and its derivatives using multilayer feedforward networks. Neural networks, 3(5):551–560, 1990.
  25. Highly accurate protein structure prediction with alphafold. Nature, 596(7873):583–589, 2021.
  26. Universal approximation with deep narrow networks. In Conference on learning theory, pages 2306–2327. PMLR, 2020.
  27. Reduction of the Hodgkin-Huxley equations to a single-variable threshold model. Neural computation, 9(5):1015–1045, 1997.
  28. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems, 25, 2012.
  29. Deep learning. nature, 521(7553):436–444, 2015.
  30. Multilayer feedforward networks with a nonpolynomial activation function can approximate any function. Neural networks, 6(6):861–867, 1993.
  31. Wolfgang Maass. Lower bounds for the computational power of networks of spiking neurons. Neural computation, 8(1):1–40, 1996.
  32. Wolfgang Maass. Networks of spiking neurons: the third generation of neural network models. Neural networks, 10(9):1659–1671, 1997.
  33. Wolfgang Maass. Noisy spiking neurons with temporal coding have more computational power than sigmoidal neurons. Advances in Neural Information Processing Systems, 9:211–217, 1997.
  34. Wolfgang Maass. To spike or not to spike: that is the question. Proceedings of the IEEE, 103(12):2219–2224, 2015.
  35. Pulsed neural networks. MIT press, 2001.
  36. On the complexity of learning for a spiking neuron. In Proceedings of the tenth annual conference on Computational learning theory, pages 54–61, 1997.
  37. On the complexity of learning for spiking neurons with temporal coding. Information and Computation, 153(1):26–46, 1999.
  38. Foundations of machine learning. MIT press, 2018.
  39. OpenAI. GPT-4 technical report. arXiv preprint 2303.0877, 2023.
  40. Deep ReLU networks and high-order finite element methods. Analysis and Applications, 18(05):715–770, 2020.
  41. Hélene Paugam-Moisy. Spiking neuron networks a survey. 2006.
  42. Topological properties of the set of functions generated by neural networks of fixed size. Foundations of computational mathematics, 21:375–444, 2021.
  43. How many neurons does it take to approximate the maximum? In Proceedings of the 2024 Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 3156–3183. SIAM, 2024.
  44. Johannes Schmidt-Hieber. Nonparametric regression using deep neural networks with relu activation function. Annals of statistics, 48(4):1875–1897, 2020.
  45. Michael Schmitt. VC dimension bounds for networks of spiking neurons. In ESANN, pages 429–434, 1999.
  46. Understanding machine learning: From theory to algorithms. Cambridge university press, 2014.
  47. Optimal approximation rate of ReLU networks in terms of width and depth. Journal de Mathématiques Pures et Appliquées, 157:101–135, 2022.
  48. High-order approximation rates for shallow neural networks with cosine and ReLUk𝑘{}^{k}start_FLOATSUPERSCRIPT italic_k end_FLOATSUPERSCRIPT activation functions. Applied and Computational Harmonic Analysis, 58:1–26, 2022.
  49. Expressivity of spiking neural networks. arXiv preprint arXiv:2308.08218, 2023.
  50. Richard B Stein. Some models of neuronal variability. Biophysical journal, 7(1):37–68, 1967.
  51. Matus Telgarsky. Representation benefits of deep feedforward networks. arXiv preprint arXiv:1509.08101, 2015.
  52. Deep learning’s diminishing returns: The cost of improvement is becoming unsustainable. ieee Spectrum, 58(10):50–55, 2021.
  53. Dmitry Yarotsky. Error bounds for approximations with deep ReLU networks. Neural Networks, 94:103–114, 2017.
  54. Understanding deep learning (still) requires rethinking generalization. Communications of the ACM, 64(3):107–115, 2021.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

Sign up for free to view the 2 tweets with 1 like about this paper.