Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
GPT-5.1
GPT-5.1 96 tok/s
Gemini 3.0 Pro 48 tok/s Pro
Gemini 2.5 Flash 155 tok/s Pro
Kimi K2 197 tok/s Pro
Claude Sonnet 4.5 36 tok/s Pro
2000 character limit reached

Parallel Hyperparameter Optimization Of Spiking Neural Network (2403.00450v1)

Published 1 Mar 2024 in cs.NE and cs.AI

Abstract: Spiking Neural Networks (SNN). SNNs are based on a more biologically inspired approach than usual artificial neural networks. Such models are characterized by complex dynamics between neurons and spikes. These are very sensitive to the hyperparameters, making their optimization challenging. To tackle hyperparameter optimization of SNNs, we initially extended the signal loss issue of SNNs to what we call silent networks. These networks fail to emit enough spikes at their outputs due to mistuned hyperparameters or architecture. Generally, search spaces are heavily restrained, sometimes even discretized, to prevent the sampling of such networks. By defining an early stopping criterion detecting silent networks and by designing specific constraints, we were able to instantiate larger and more flexible search spaces. We applied a constrained Bayesian optimization technique, which was asynchronously parallelized, as the evaluation time of a SNN is highly stochastic. Large-scale experiments were carried-out on a multi-GPU Petascale architecture. By leveraging silent networks, results show an acceleration of the search, while maintaining good performances of both the optimization algorithm and the best solution obtained. We were able to apply our methodology to two popular training algorithms, known as spike timing dependent plasticity and surrogate gradient. Early detection allowed us to prevent worthless and costly computation, directing the search toward promising hyperparameter combinations. Our methodology could be applied to multi-objective problems, where the spiking activity is often minimized to reduce the energy consumption. In this scenario, it becomes essential to find the delicate frontier between low-spiking and silent networks. Finally, our approach may have implications for neural architecture search, particularly in defining suitable spiking architectures.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (40)
  1. doi:10.1109/JSSC.2013.2259038. URL http://ieeexplore.ieee.org/document/6515159/
  2. doi:10.1109/MM.2018.112130359. URL https://ieeexplore.ieee.org/document/8259423/
  3. doi:10.1016/S0893-6080(97)00011-7. URL https://linkinghub.elsevier.com/retrieve/pii/S0893608097000117
  4. doi:10.1038/s43588-021-00184-y. URL https://www.nature.com/articles/s43588-021-00184-y
  5. doi:10.1109/IJCNN48605.2020.9207239. URL https://ieeexplore.ieee.org/document/9207239/
  6. doi:10.1002/widm.1484. URL https://wires.onlinelibrary.wiley.com/doi/10.1002/widm.1484
  7. doi:10.1145/3439730. URL https://dl.acm.org/doi/10.1145/3439730
  8. doi:10.1109/BigData47090.2019.9006383. URL https://ieeexplore.ieee.org/document/9006383/
  9. doi:10.3389/fnins.2020.00667. URL https://www.frontiersin.org/article/10.3389/fnins.2020.00667/full
  10. doi:10.1109/GRID.2005.1542730. URL http://ieeexplore.ieee.org/document/1542730/
  11. doi:10.1016/j.neucom.2020.07.061. URL https://linkinghub.elsevier.com/retrieve/pii/S0925231220311693
  12. doi:10.1109/IJCNN48605.2020.9206681. URL https://ieeexplore.ieee.org/document/9206681/
  13. doi:10.1016/j.neucom.2013.06.052. URL https://linkinghub.elsevier.com/retrieve/pii/S0925231214003452
  14. doi:10.1109/TNN.2003.820440. URL http://ieeexplore.ieee.org/document/1257420/
  15. doi:10.1113/jphysiol.1952.sp004717. URL https://onlinelibrary.wiley.com/doi/10.1113/jphysiol.1952.sp004717
  16. doi:10.1007/s00354-022-00165-3. URL https://link.springer.com/10.1007/s00354-022-00165-3
  17. doi:10.23919/DATE56975.2023.10136914. URL https://ieeexplore.ieee.org/document/10136914/
  18. doi:10.1109/TNNLS.2021.3055421. URL https://ieeexplore.ieee.org/document/9353400/
  19. doi:10.1007/s00521-021-06824-8. URL https://link.springer.com/10.1007/s00521-021-06824-8
  20. doi:10.3389/fncom.2015.00099. URL http://journal.frontiersin.org/Article/10.3389/fncom.2015.00099/abstract
  21. doi:10.1145/2950067.2950090.
  22. doi:10.1109/TNNLS.2019.2941506. URL https://ieeexplore.ieee.org/document/8867860/
  23. doi:10.7554/eLife.47314. URL https://elifesciences.org/articles/47314
  24. doi:10.1109/IJCNN48605.2020.9207474. URL https://ieeexplore.ieee.org/document/9207474/
  25. doi:10.1088/2634-4386/ac8bef. URL https://iopscience.iop.org/article/10.1088/2634-4386/ac8bef
  26. doi:10.1007/s11042-023-16344-3. URL https://link.springer.com/10.1007/s11042-023-16344-3
  27. doi:10.1145/3477145.3477160. URL https://dl.acm.org/doi/10.1145/3477145.3477160
  28. doi:10.1016/j.neunet.2018.03.019. URL https://linkinghub.elsevier.com/retrieve/pii/S0893608018301126
  29. doi:10.1145/3381755.3381758. URL https://dl.acm.org/doi/10.1145/3381755.3381758
  30. doi:10.1609/aaai.v35i12.17320. URL https://ojs.aaai.org/index.php/AAAI/article/view/17320
  31. doi:10.1016/j.neunet.2021.09.022. URL https://linkinghub.elsevier.com/retrieve/pii/S0893608021003841
  32. doi:10.1109/ACCESS.2020.3047071. URL https://ieeexplore.ieee.org/document/9306772/
  33. doi:10.1109/UEMCON47517.2019.8992983. URL https://ieeexplore.ieee.org/document/8992983/
  34. doi:10.1109/IJCNN.2018.8489673. URL https://ieeexplore.ieee.org/document/8489673/
  35. doi:10.3389/fninf.2018.00089. URL https://www.frontiersin.org/article/10.3389/fninf.2018.00089/full
  36. doi:10.1016/j.neucom.2020.04.079. URL https://linkinghub.elsevier.com/retrieve/pii/S0925231220306640
  37. doi:10.1109/IJCNN.2018.8489684. URL https://ieeexplore.ieee.org/document/8489684/
  38. doi:10.3389/fnins.2018.00774. URL https://www.frontiersin.org/article/10.3389/fnins.2018.00774/full
  39. doi:10.1109/CVPR.2017.781. URL https://ieeexplore.ieee.org/document/8100264/
  40. doi:10.5281/zenodo.5079802. URL https://doi.org/10.5281/zenodo.5079802
Citations (1)

Summary

  • The paper introduces a parallel Bayesian algorithm that efficiently optimizes hyperparameters of SNNs by eliminating silent networks early.
  • It employs a spike-based early stopping criterion to allocate computational resources to viable, dynamic spiking networks, reducing overhead.
  • Experiments on STDP and SLAYER training across MNIST and DVS128 show significant improvements in resource utilization and scalability.

Enhancing the Efficiency of Spiking Neural Networks Through Parallel Hyperparameter Optimization

Introduction

The optimization of Spiking Neural Networks (SNNs) poses a unique set of challenges not encountered in the training of traditional Artificial Neural Networks. This paper explores a novel approach to hyperparameter optimization (HPO) in SNNs that leverages parallel computation and constrained Bayesian optimization. The focus is on addressing the challenges posed by "silent networks" - a scenario where networks show insufficient spiking activity, making them infeasible solutions for tasks requiring dynamic spiking responses.

Methodology

The methodology centers around the development of a scalable, constrained Bayesian-based optimization algorithm, specifically designed to mitigate the sampling of silent networks within high-dimensional search spaces. The optimization process is enhanced by a spike-based early stopping criterion, which swiftly identifies and discards silent networks, thereby speeding up the optimization cycle. Furthermore, this approach is parallelized through asynchronous computation on multi-GPU Petascale architectures, ensuring scalability and efficiency.

Experiments

Experiments are carried out using two primary training algorithms: Spike Timing Dependent Plasticity (STDP) and a surrogate gradient method (SLAYER), across different datasets (MNIST encoded in Poisson spikes and DVS128 Gesture). The paper rigorously tests the optimization algorithm across a variety of SNN configurations, assessing its effectiveness in navigating through high-dimensional search spaces while avoiding the computation of silent networks.

Results

The optimization algorithm demonstrates a remarkable ability to identify and focus on viable spiking networks, significantly mitigating the computational overhead associated with silent networks. For instance, despite a considerable portion of the networks being stopped due to insufficient spiking activity (73% in one experiment), these networks only consumed 36% of the total computational resources. This indicates the efficiency of the spike-based early stopping criterion in conjunction with the optimization algorithm in preserving computational resources for more promising network configurations.

Implications

The findings have several practical and theoretical implications:

  • Scalability: The approach effectively handles high-dimensional search spaces, a crucial consideration for the complex architectures typical of SNNs.
  • Efficiency: By focusing computational resources on feasible solutions, the methodology significantly improves the efficiency of the HPO process.
  • Generalization: The methodology is adaptable to both popular families of SNN training algorithms, showcasing its versatility.
  • Future Research Directions: The paper opens avenues for further investigation, particularly around multi-objective optimization and its applicability in Neural Architecture Search (NAS) for SNNs.

Conclusion

This paper presents a pioneering approach to the hyperparameter optimization of Spiking Neural Networks by integrating a novel early stopping criterion and constrained Bayesian optimization within a parallel computational framework. The methodology not only addresses the specific challenge of silent networks but also sets a foundation for future explorations into efficient and scalable HPO strategies for SNNs.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 1 tweet and received 1 like.

Upgrade to Pro to view all of the tweets about this paper: