Complemented Ternary Spiking Neuron
- CTSN is a spiking neuron model that integrates ternary signaling with a learnable, non-resetting memory to mitigate iterative information loss.
- It enhances temporal gradient flow and network heterogeneity, leading to improved performance on both static image and neuromorphic benchmarks.
- The model supports efficient hardware implementation via superconducting circuits, achieving high-speed, energy-efficient computation with state-of-the-art accuracy.
A Complemented Ternary Spiking Neuron (CTSN) is a model of spiking neuron that integrates ternary-valued signaling with a learnable memory mechanism, designed to improve information retention, temporal gradient flow, and network heterogeneity within both algorithmic and hardware SNN implementations. This model advances on prior spiking neuron architectures by embedding adaptive, non-resetting memory and supporting excitatory-inhibitory balance, while offering straightforward mapping to high-speed hardware such as superconductor-based circuits. CTSN models have been demonstrated to achieve state-of-the-art performance on a variety of neuromorphic and static image benchmarks and form the computational primitives of recent ultra-fast, cryogenic SNN accelerators (Zhang et al., 22 Jan 2026, Karamuftuoglu et al., 2024).
1. Mathematical Definition and Dynamics
The baseline ternary spiking neuron model, as formalized by Guo et al. (2024), consists of leaky membrane updates, hard resets on spike emission, and ternary output quantization:
- Membrane update:
- Ternary spike:
The CTSN augments this framework with a learnable, non-resetting complemental term , which preserves partial state across spikes:
- Integrated update:
\begin{align*} u{(l)}(t) &= \tau\,\tilde{u}{(l)}(t-1)\,(1 - |o{(l)}(t-1)|)\ \tilde{u}{(l)}(t) &= h{(l)}(t) + x{(l)}(t)\ o{(l)}(t) &= \Theta(\tilde{u}{(l)}(t), V_{\rm th}) \end{align*}
- Complemental memory dynamics:
For static image inputs: For neuromorphic/dynamic inputs: with and typically initialized to 0.5.
The presence of endows the neuron with heterogeneity and partial memory, counteracting iterative loss due to reset and supporting richer temporal dynamics (Zhang et al., 22 Jan 2026).
2. Biological and Algorithmic Motivation
Vanilla ternary spiking neurons experience full memory loss after each spike due to the hard reset, precluding any influence of pre-spike states on future output—this is described as iterative information loss. Additionally, in backpropagation-through-time (BPTT), two compounding effects undermine stable learning: rapid attenuation due to repeated factors of small , and abrupt gradient vanishing whenever large spikes occur (). The complemental term bypasses hard resets, enabling ongoing integration of past states and introducing dynamic heterogeneity at the neuron level. In gradient-based learning, supplies alternative gradient pathways (), preventing rapid decay and supporting reliable temporal credit assignment (Zhang et al., 22 Jan 2026).
3. Temporal Membrane Potential Regularization (TMPR)
CTSNs are typically trained jointly with a TMPR loss, designed to regularize the distribution and magnitude of time-evolving membrane potentials:
- Loss function:
, with
where is batch size, the number of neurons in layer , the time horizon, the number of layers, and a small hyperparameter.
- The $1/t$ factor exerts strong regularization early in the sequence to prevent premature over-activation and decays temporally to favor use of later memory.
- TMPR directly creates spatiotemporal shortcut gradients (), augmenting the gradient signal available to earlier time steps and deep layers.
This approach demonstrably stabilizes training and mitigates the temporal vanishing gradient phenomenon in recurrent SNNs (Zhang et al., 22 Jan 2026).
4. Hardware Realization and Ternary Superconducting Circuits
A complementary hardware instantiation of CTSN has been engineered using superconductor electronics, leveraging single flux quantum (SFQ) logic and Josephson Junctions (JJ) (Karamuftuoglu et al., 2024):
- Each synapse is implemented as ternary (), configured via mutual inductive coupling to the neuron loop.
- Input spikes are encoded as SFQ voltage pulses; excitatory/inhibitory weighting is enacted by routing these pulses through positive/negative inductive loops.
- The neuron loop contains two output JJs (JJP for positive, JJN for negative). Whichever output threshold is crossed first fires the corresponding output.
- Leaky integration is realized via resistive branches in each dendrite, establishing the time constant .
This hardware CTSN supports high fan-in architectures, asynchronous firing, and dual-polarity outputs, capturing the complemented ternary operation at the physical level (Karamuftuoglu et al., 2024).
| Feature | Algorithmic CTSN (Zhang et al., 22 Jan 2026) | Superconductor CTSN (Karamuftuoglu et al., 2024) |
|---|---|---|
| Memory mechanism | Learnable complemental term | Persistent loop current, no hard reset |
| Ternary signaling | output | SFQ pulses, positive/negative coupling |
| Reset behavior | Partial; never resets | Loop reset only when either JJ fires |
5. Pseudocode and Training Integration
Forward and backward passes for CTSN employ the following algorithmic structure:
- For each time step:
- Update ,
- Update using ,
- Compute ,
- Output ,
- Post-spike hard reset: .
Surrogate gradients are applied to the non-differentiable using a rectangular window of width . Training is performed using SGD with standard spiking data augmentation and weight decay. For hardware-oriented SNNs, quantized weights are pruned and mapped directly to SFQ-coupled circuits after training (using straight-through estimators and in-situ masking).
6. Empirical Results and Performance Characteristics
Across static image and neuromorphic datasets, CTSN consistently improves over baseline and SOTA ternary SNNs. Key results:
- Ablation (ResNet20):
- CIFAR-100, : Baseline 73.95%, +CTSN 74.48%, +CTSN+TMPR 74.82%
- CIFAR10-DVS, : Baseline 80.30%, +CTSN 80.57%, +CTSN+TMPR 81.23%
- Full-network SOTA Scores:
- CIFAR-10, ResNet19: 96.28% (ternary) → 96.46% (CTSN+TMPR)
- CIFAR-100, ResNet19: 79.68% → 81.19%
- ImageNet-100, SEW-ResNet34: 83.27% (prior best) → 85.06% (CTSN+TMPR)
- Neuromorphic (VGGSNN, CIFAR10-DVS): 83.20% (CTSN+TMPR open-source reimpl.)
- Temporal ablation: For short sequence lengths (), CTSN yields negligible difference. Gains increase with , supporting the interpretation that improves long-term credit assignment by maintaining informative state.
- Hardware platform: On MNIST, a SFQ-based CTSN SNN (784–128–96–96–10) achieves 96.1% accuracy with 64-synapse pruning, 8.92 GHz throughput, and 1.5 nJ per inference (including 4 K cooling). Ternary pruning reduces static power by 82.7% with a 0.97% accuracy drop (Karamuftuoglu et al., 2024).
7. Architectural and Hyperparameter Considerations
- Surrogate gradient: rectangular window, width
- Decay constant: (all experiments)
- Training: SGD, momentum 0.9, cosine-annealed LR, batch size 64, 300 epochs
- Weight decay: (static), (DVS)
- Input encoding: Direct multi-step encoding for static, 1-bit Poisson for hardware SNNs
- Data augmentation: Standard flips/crops, AutoAugment, Cutout (static), random roll+flip (DVS)
CTSNs require three additional per-neuron scalars (), imposing a minimal parameter overhead while furnishing substantial gains in memory retention, gradient flow, and task performance across diverse SNN architectures (Zhang et al., 22 Jan 2026, Karamuftuoglu et al., 2024).