Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
92 tokens/sec
Gemini 2.5 Pro Premium
51 tokens/sec
GPT-5 Medium
24 tokens/sec
GPT-5 High Premium
17 tokens/sec
GPT-4o
97 tokens/sec
DeepSeek R1 via Azure Premium
92 tokens/sec
GPT OSS 120B via Groq Premium
458 tokens/sec
Kimi K2 via Groq Premium
222 tokens/sec
2000 character limit reached

Low-Bit Data Processing Using Multiple-Output Spiking Neurons with Non-linear Reset Feedback (2508.06292v1)

Published 8 Aug 2025 in cs.LG

Abstract: Neuromorphic computing is an emerging technology enabling low-latency and energy-efficient signal processing. A key algorithmic tool in neuromorphic computing is spiking neural networks (SNNs). SNNs are biologically inspired neural networks which utilize stateful neurons, and provide low-bit data processing by encoding and decoding information using spikes. Similar to SNNs, deep state-space models (SSMs) utilize stateful building blocks. However, deep SSMs, which recently achieved competitive performance in various temporal modeling tasks, are typically designed with high-precision activation functions and no reset mechanisms. To bridge the gains offered by SNNs and the recent deep SSM models, we propose a novel multiple-output spiking neuron model that combines a linear, general SSM state transition with a non-linear feedback mechanism through reset. Compared to the existing neuron models for SNNs, our proposed model clearly conceptualizes the differences between the spiking function, the reset condition and the reset action. The experimental results on various tasks, i.e., a keyword spotting task, an event-based vision task and a sequential pattern recognition task, show that our proposed model achieves performance comparable to existing benchmarks in the SNN literature. Our results illustrate how the proposed reset mechanism can overcome instability and enable learning even when the linear part of neuron dynamics is unstable, allowing us to go beyond the strictly enforced stability of linear dynamics in recent deep SSM models.

Summary

  • The paper presents a novel multi-output spiking neuron architecture that integrates a non-linear reset feedback to stabilize learning under low-bit conditions.
  • The method combines stateful SNN dynamics with deep SSM computational strengths through a learnable reset mechanism triggered when thresholds are exceeded.
  • Experimental results indicate competitive performance in tasks like keyword spotting and event-based vision, with minimal computational overhead suitable for edge deployment.

Low-Bit Data Processing Using Multiple-Output Spiking Neurons with Non-linear Reset Feedback

Introduction

Neuromorphic computing represents an innovative approach to achieving low-latency and energy-efficient signal processing through the application of Spiking Neural Networks (SNNs). These networks, inspired by biological neurons, utilize stateful units to process low-bit data effectively. While SNNs traditionally employ biologically plausible models such as Integrate-and-Fire (IF) neurons, this paper introduces a novel approach integrating aspects of both SNNs and Deep State-Space Models (SSMs). SSMs are typically designed with high-precision activation functions and lack reset mechanisms. The proposed model introduces a multiple-output spiking neuron architecture that combines linear state transitions with a non-linear feedback mechanism through a reset. This enables the exploitation of the advantages offered by the stateful dynamics of SNNs with the computational prowess of deep SSMs.

Proposed Neuron Model

The proposed neuron model leverages a multiple-output framework with a general state transition matrix and a reset mechanism defined by learnable parameters. This contrasts with the typical single-output model prevalent in SNN literature. By embracing a complex-valued diagonal state transition, the neuron model allows for enriched dynamics. Specifically, the reset condition and action are separated from conventional spiking behavior, offering a unique mechanism where the reset is triggered not solely by spikes but by exceeding a specific threshold. This mechanism can facilitate learning even when the neuron dynamics are unstable, which is generally constrained in typical SSM models. Figure 1

Figure 1: Illustration of dynamics of a neuron with multiple-outputs and a general reset mechanism.

Experimental Results

The experimental evaluation consists of tasks such as keyword spotting, event-based vision, and sequential pattern recognition. The neuron model demonstrates comparable performance against existing benchmarks in the SNN literature. Notably, the reset mechanism proved advantageous in overcoming instability within neuron dynamics, leading to robust learning processes even under conditions where high precision is not maintained. Figure 2

Figure 2

Figure 2: Illustration of adLIF neuron reset condition with v1>θv_1>\theta.

The results consistently show that the proposed architecture holds promise in various data modalities, emphasizing its applicability across different domains. The versatility is further demonstrated by its ability to generalize across multilingual spoken word recognition and gesture-based data processing.

Computational Considerations

Implementing the proposed model requires acknowledging the computational overhead introduced primarily by the reset mechanism. This overhead is minimal, consisting of three additional learnable parameters per neuron, and slightly increased computational operations without significantly affecting the overall efficiency. The architecture ensures scalability and deployability, suitable for edge applications where power and latency constraints are paramount.

Conclusion

The paper introduces an innovative multiple-output spiking neuron model that bridges the gap between traditional SNNs and deep SSM architectures. It highlights the potential benefits of integrating non-linear reset mechanisms to enhance learning stability, particularly within unstable neuron dynamics. Such integration opens pathways for more expressive and biologically inspired SSM-based architectures, potentially enriching both theoretical understanding and practical implementation in neuromorphic computing scenarios. Future investigations may explore refining learnable reset mechanisms and exploring real-time deployment in neuromorphic hardware platforms.