Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
81 tokens/sec
Gemini 2.5 Pro Premium
47 tokens/sec
GPT-5 Medium
22 tokens/sec
GPT-5 High Premium
20 tokens/sec
GPT-4o
88 tokens/sec
DeepSeek R1 via Azure Premium
79 tokens/sec
GPT OSS 120B via Groq Premium
459 tokens/sec
Kimi K2 via Groq Premium
192 tokens/sec
2000 character limit reached

Learning fast changing slow in spiking neural networks (2402.10069v2)

Published 25 Jan 2024 in cs.NE and cs.LG

Abstract: Reinforcement learning (RL) faces substantial challenges when applied to real-life problems, primarily stemming from the scarcity of available data due to limited interactions with the environment. This limitation is exacerbated by the fact that RL often demands a considerable volume of data for effective learning. The complexity escalates further when implementing RL in recurrent spiking networks, where inherent noise introduced by spikes adds a layer of difficulty. Life-long learning machines must inherently resolve the plasticity-stability paradox. Striking a balance between acquiring new knowledge and maintaining stability is crucial for artificial agents. To address this challenge, we draw inspiration from machine learning technology and introduce a biologically plausible implementation of proximal policy optimization, referred to as lf-cs (learning fast changing slow). Our approach results in two notable advancements: firstly, the capacity to assimilate new information into a new policy without requiring alterations to the current policy; and secondly, the capability to replay experiences without experiencing policy divergence. Furthermore, when contrasted with other experience replay (ER) techniques, our method demonstrates the added advantage of being computationally efficient in an online setting. We demonstrate that the proposed methodology enhances the efficiency of learning, showcasing its potential impact on neuromorphic and real-world applications.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (22)
  1. Stochastic variational learning in recurrent spiking networks. Frontiers in computational neuroscience, 8:38, 2014.
  2. Predicting non-linear dynamics by stable local learning in a recurrent spiking neural network. Elife, 6:e28295, 2017.
  3. A solution to the learning dilemma for recurrent networks of spiking neurons. Nature communications, 11(1):1–15, 2020.
  4. The remarkable robustness of surrogate gradient learning for instilling complex function in spiking neural networks. Neural computation, 33(4):899–925, 2021.
  5. Error-based or target-based? a unified framework for learning in recurrent spiking networks. PLoS computational biology, 18(6):e1010221, 2022.
  6. Target spike patterns enable efficient and biologically plausible learning for complex temporal tasks. PloS one, 16(2):e0247014, 2021.
  7. full-force: A target-based method for training recurrent networks. PloS one, 13(2):e0191527, 2018.
  8. Alessandro Ingrosso and LF Abbott. Training dynamically balanced excitatory-inhibitory networks. PloS one, 14(8):e0220547, 2019.
  9. Sleep-like slow oscillations improve visual classification through synaptic homeostasis and memory association in a thalamo-cortical model. Scientific Reports, 9(1):1–11, 2019.
  10. Dendrites help mitigate the plasticity-stability dilemma. Scientific Reports, 13(1):6543, 2023.
  11. Beyond spiking networks: The computational advantages of dendritic amplification and input segregation. Proceedings of the National Academy of Sciences, 120(49):e2220743120, 2023.
  12. Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347, 2017.
  13. Sample efficient actor-critic with experience replay. arXiv preprint arXiv:1611.01224, 2016.
  14. Trust region policy optimization. In International conference on machine learning, pages 1889–1897. PMLR, 2015.
  15. Asynchronous methods for deep reinforcement learning. In International conference on machine learning, pages 1928–1937. PMLR, 2016.
  16. Openai gym, 2016.
  17. Reinforcement learning: An introduction. MIT press, 2018.
  18. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  19. Evolving connectivity for recurrent spiking neural networks. arXiv preprint arXiv:2305.17650, 2023.
  20. An event-driven recurrent spiking neural network architecture for efficient inference on fpga. In Proceedings of the International Conference on Neuromorphic Systems 2022, pages 1–8, 2022.
  21. Combining stdp and binary networks for reinforcement learning from images and sparse rewards. Neural Networks, 144:496–506, 2021.
  22. Brain-inspired neural circuit evolution for spiking neural networks.ÍÍ Proceedings of the National Academy of Sciences, 120(39):e2218173120, 2023.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com