Papers
Topics
Authors
Recent
Search
2000 character limit reached

Complemented Ternary Spiking Neuron

Updated 29 January 2026
  • CTSN is a spiking neuron model that integrates ternary signaling with a learnable, non-resetting memory to mitigate iterative information loss.
  • It enhances temporal gradient flow and network heterogeneity, leading to improved performance on both static image and neuromorphic benchmarks.
  • The model supports efficient hardware implementation via superconducting circuits, achieving high-speed, energy-efficient computation with state-of-the-art accuracy.

A Complemented Ternary Spiking Neuron (CTSN) is a model of spiking neuron that integrates ternary-valued signaling with a learnable memory mechanism, designed to improve information retention, temporal gradient flow, and network heterogeneity within both algorithmic and hardware SNN implementations. This model advances on prior spiking neuron architectures by embedding adaptive, non-resetting memory and supporting excitatory-inhibitory balance, while offering straightforward mapping to high-speed hardware such as superconductor-based circuits. CTSN models have been demonstrated to achieve state-of-the-art performance on a variety of neuromorphic and static image benchmarks and form the computational primitives of recent ultra-fast, cryogenic SNN accelerators (Zhang et al., 22 Jan 2026, Karamuftuoglu et al., 2024).

1. Mathematical Definition and Dynamics

The baseline ternary spiking neuron model, as formalized by Guo et al. (2024), consists of leaky membrane updates, hard resets on spike emission, and ternary output quantization:

  • Membrane update:

u(l)(t)=τu(l)(t1)(1o(l)(t1))+x(l)(t),x(l)(t)=W(l)o(l1)(t)+b(l)u^{(l)}(t) = \tau\,u^{(l)}(t\mathbin{-}1)\,(1 - |o^{(l)}(t\mathbin{-}1)|) + x^{(l)}(t), \quad x^{(l)}(t) = W^{(l)} o^{(l-1)}(t) + b^{(l)}

  • Ternary spike:

o(l)(t)=Θ(u(l)(t),Vth)={+1if uVth 1if uVth 0otherwiseo^{(l)}(t) = \Theta(u^{(l)}(t), V_{\rm th}) = \begin{cases} +1 & \text{if } u \geq V_{\rm th} \ -1 & \text{if } u \leq -V_{\rm th} \ 0 & \text{otherwise} \end{cases}

The CTSN augments this framework with a learnable, non-resetting complemental term h(l)(t)h^{(l)}(t), which preserves partial state across spikes:

  • Integrated update:

\begin{align*} u{(l)}(t) &= \tau\,\tilde{u}{(l)}(t-1)\,(1 - |o{(l)}(t-1)|)\ \tilde{u}{(l)}(t) &= h{(l)}(t) + x{(l)}(t)\ o{(l)}(t) &= \Theta(\tilde{u}{(l)}(t), V_{\rm th}) \end{align*}

  • Complemental memory dynamics:

For static image inputs: h(l)(t)=αReLU(h(l)(t1))+β(ReLU(h(l)(t1)))+γu(l)(t)h^{(l)}(t) = \alpha \operatorname{ReLU}(h^{(l)}(t-1)) + \beta(-\operatorname{ReLU}(-h^{(l)}(t-1))) + \gamma u^{(l)}(t) For neuromorphic/dynamic inputs: h(l)(t)=αh(l)(t1)+βReLU(u(l)(t))+γ(ReLU(u(l)(t)))h^{(l)}(t) = \alpha h^{(l)}(t-1) + \beta \operatorname{ReLU}(u^{(l)}(t)) + \gamma(-\operatorname{ReLU}(-u^{(l)}(t))) with α,β,γ=σ(ωa),σ(ωb),σ(ωc)(0,1)\alpha, \beta, \gamma = \sigma(\omega_a), \sigma(\omega_b), \sigma(\omega_c) \in (0,1) and typically initialized to 0.5.

The presence of h(l)(t)h^{(l)}(t) endows the neuron with heterogeneity and partial memory, counteracting iterative loss due to reset and supporting richer temporal dynamics (Zhang et al., 22 Jan 2026).

2. Biological and Algorithmic Motivation

Vanilla ternary spiking neurons experience full memory loss after each spike due to the hard reset, precluding any influence of pre-spike states on future output—this is described as iterative information loss. Additionally, in backpropagation-through-time (BPTT), two compounding effects undermine stable learning: rapid attenuation due to repeated factors of small τ<1\tau < 1, and abrupt gradient vanishing whenever large spikes occur (u>1.5Vth|u| > 1.5 V_{\rm th}). The complemental term h(t)h(t) bypasses hard resets, enabling ongoing integration of past states and introducing dynamic heterogeneity at the neuron level. In gradient-based learning, h(t)h(t) supplies alternative gradient pathways (h(t+1)/h(t)>0\partial h(t+1)/\partial h(t) > 0), preventing rapid decay and supporting reliable temporal credit assignment (Zhang et al., 22 Jan 2026).

3. Temporal Membrane Potential Regularization (TMPR)

CTSNs are typically trained jointly with a TMPR loss, designed to regularize the distribution and magnitude of time-evolving membrane potentials:

  • Loss function:

L=LCE+LTMPR\mathcal{L} = \mathcal{L}_{\rm CE} + \mathcal{L}_{\rm TMPR}, with

LTMPR=1TLt=1Tλtl=1Lu~(l)(t)22/(BD(l))\mathcal{L}_{\rm TMPR} = \frac{1}{TL}\sum_{t=1}^T\frac{\lambda}{t}\sum_{l=1}^L \|\tilde{u}^{(l)}(t)\|_2^2/(B D^{(l)})

where BB is batch size, D(l)D^{(l)} the number of neurons in layer ll, TT the time horizon, LL the number of layers, and λ\lambda a small hyperparameter.

  • The $1/t$ factor exerts strong regularization early in the sequence to prevent premature over-activation and decays temporally to favor use of later memory.
  • TMPR directly creates spatiotemporal shortcut gradients (L/u~(l)(t)\partial \mathcal{L}/\partial \tilde{u}^{(l)}(t)), augmenting the gradient signal available to earlier time steps and deep layers.

This approach demonstrably stabilizes training and mitigates the temporal vanishing gradient phenomenon in recurrent SNNs (Zhang et al., 22 Jan 2026).

4. Hardware Realization and Ternary Superconducting Circuits

A complementary hardware instantiation of CTSN has been engineered using superconductor electronics, leveraging single flux quantum (SFQ) logic and Josephson Junctions (JJ) (Karamuftuoglu et al., 2024):

  • Each synapse is implemented as ternary (w={1,0,+1}w = \{-1, 0, +1\}), configured via mutual inductive coupling to the neuron loop.
  • Input spikes are encoded as SFQ voltage pulses; excitatory/inhibitory weighting is enacted by routing these pulses through positive/negative inductive loops.
  • The neuron loop contains two output JJs (JJP for positive, JJN for negative). Whichever output threshold is crossed first fires the corresponding output.
  • Leaky integration is realized via resistive branches in each dendrite, establishing the time constant τ=Ltotal/Rtotal\tau = L_{\text{total}}/R_{\text{total}}.

This hardware CTSN supports high fan-in architectures, asynchronous firing, and dual-polarity outputs, capturing the complemented ternary operation at the physical level (Karamuftuoglu et al., 2024).

Feature Algorithmic CTSN (Zhang et al., 22 Jan 2026) Superconductor CTSN (Karamuftuoglu et al., 2024)
Memory mechanism Learnable complemental term h(t)h(t) Persistent loop current, no hard reset
Ternary signaling {1,0,+1}\{-1, 0, +1\} output SFQ pulses, positive/negative coupling
Reset behavior Partial; h(t)h(t) never resets Loop reset only when either JJ fires

5. Pseudocode and Training Integration

Forward and backward passes for CTSN employ the following algorithmic structure:

  • For each time step:
    1. Update u=τu~(1ot1)u = \tau\,\tilde{u}\,(1 - |o_{t-1}|),
    2. Update hh using h=G(h,u;α,β,γ)h = \mathcal{G}(h, u; \alpha, \beta, \gamma),
    3. Compute u~=h+Wol1(t)+b\tilde{u} = h + W o^{l-1}(t) + b,
    4. Output o(t)=Θ(u~,Vth)o(t) = \Theta(\tilde{u}, V_{\rm th}),
    5. Post-spike hard reset: u~u~(1o(t))\tilde{u} \leftarrow \tilde{u}\,(1 - |o(t)|).

Surrogate gradients are applied to the non-differentiable Θ\Theta using a rectangular window of width a=0.5a=0.5. Training is performed using SGD with standard spiking data augmentation and weight decay. For hardware-oriented SNNs, quantized weights are pruned and mapped directly to SFQ-coupled circuits after training (using straight-through estimators and in-situ masking).

6. Empirical Results and Performance Characteristics

Across static image and neuromorphic datasets, CTSN consistently improves over baseline and SOTA ternary SNNs. Key results:

  • Ablation (ResNet20):
    • CIFAR-100, T=4T=4: Baseline 73.95%, +CTSN 74.48%, +CTSN+TMPR 74.82%
    • CIFAR10-DVS, T=10T=10: Baseline 80.30%, +CTSN 80.57%, +CTSN+TMPR 81.23%
  • Full-network SOTA Scores:
    • CIFAR-10, ResNet19: 96.28% (ternary) → 96.46% (CTSN+TMPR)
    • CIFAR-100, ResNet19: 79.68% → 81.19%
    • ImageNet-100, SEW-ResNet34: 83.27% (prior best) → 85.06% (CTSN+TMPR)
    • Neuromorphic (VGGSNN, CIFAR10-DVS): 83.20% (CTSN+TMPR open-source reimpl.)
  • Temporal ablation: For short sequence lengths (T2T \leq 2), CTSN yields negligible difference. Gains increase with TT, supporting the interpretation that h(t)h(t) improves long-term credit assignment by maintaining informative state.
  • Hardware platform: On MNIST, a SFQ-based CTSN SNN (784–128–96–96–10) achieves 96.1% accuracy with 64-synapse pruning, 8.92 GHz throughput, and \sim1.5 nJ per inference (including 4 K cooling). Ternary pruning reduces static power by 82.7% with a 0.97% accuracy drop (Karamuftuoglu et al., 2024).

7. Architectural and Hyperparameter Considerations

  • Surrogate gradient: rectangular window, width a=Vth=0.5a = V_{\rm th} = 0.5
  • Decay constant: τ=0.25\tau = 0.25 (all experiments)
  • Training: SGD, momentum 0.9, cosine-annealed LR, batch size 64, 300 epochs
  • Weight decay: 1×1041 \times 10^{-4} (static), 5×1045 \times 10^{-4} (DVS)
  • Input encoding: Direct multi-step encoding for static, 1-bit Poisson for hardware SNNs
  • Data augmentation: Standard flips/crops, AutoAugment, Cutout (static), random roll+flip (DVS)

CTSNs require three additional per-neuron scalars (α,β,γ\alpha, \beta, \gamma), imposing a minimal parameter overhead while furnishing substantial gains in memory retention, gradient flow, and task performance across diverse SNN architectures (Zhang et al., 22 Jan 2026, Karamuftuoglu et al., 2024).

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Complemented Ternary Spiking Neuron (CTSN).