Papers
Topics
Authors
Recent
Search
2000 character limit reached

Self-Prediction LIF Neuron

Updated 2 February 2026
  • The paper introduces the S-LIF neuron, which augments traditional LIF dynamics with an internal prediction current from input-output history.
  • It employs a surrogate gradient method to address spike non-differentiability and vanishing gradients, boosting training efficiency and stability.
  • Empirical results demonstrate consistent accuracy gains across image, sequential, and reinforcement learning tasks while maintaining biological plausibility.

The Self-Prediction Enhanced LIF (S-LIF) neuron is a modification of the canonical leaky integrate-and-fire (LIF) spiking neuron that incorporates predictive coding principles by generating an internal prediction current from its input-output history. This construct augments the membrane dynamics, providing new gradient flow pathways that mitigate spike non-differentiability and vanishing gradients, yielding improvements in stability, training efficiency, and biological plausibility. The method demonstrates consistent performance enhancements across various network architectures, neuron types, and domains, including image and sequential classification as well as reinforcement learning (Huang et al., 29 Jan 2026).

1. Standard LIF Neuron Dynamics

The LIF neuron describes membrane potential evolution under synaptic input, leakage, thresholding, and reset. The continuous-time dynamics are formalized by

τmdu(t)dt=u(t)+Isyn(t)\tau_m\,\frac{du(t)}{dt} = -u(t) + I_{\rm syn}(t)

where u(t)u(t) is the membrane potential, τm\tau_m is the membrane time constant, and Isyn(t)I_{\rm syn}(t) is the synaptic input current. A spike is emitted when u(t)u(t) crosses the threshold VthV_{\rm th}, producing output s(t)=1s(t) = 1 and resetting u(t)u(t) to VresetV_{\rm reset}. Otherwise, s(t)=0s(t) = 0.

2. Internal Prediction Current Formulation

The S-LIF neuron augments total input current with a learned prediction current Ipred(t)I_{\rm pred}(t), generated online from the neuron's own input–output history. The prediction-current state for layer ll at time tt is

Ipredl[t]=mpl[t1]I_{\rm pred}^l[t] = \bm{m}_p^l[t-1]

where mpl[t]RN\bm{m}_p^l[t] \in \mathbb{R}^N is the internal prediction-current buffer. The update rule is

mpl[t]=(1τpl)mpl[t1]+τpl(xl1[t]sl[t]τl)\bm{m}_p^l[t] = (1-\tau_p^l)\,\bm{m}_p^l[t-1] + \tau_p^l\Big(\bm{x}^{l-1}[t] - \frac{\bm{s}^l[t]}{\tau^l}\Big)

with xl1[t]=Wlsl1[t]\bm{x}^{l-1}[t] = \bm{W}^l\,\bm{s}^{l-1}[t] as raw synaptic input, sl[t]\bm{s}^l[t] as the spike vector, τpl(0,1)\tau_p^l \in (0,1) as a learnable update rate, and τl\tau^l as the layer's membrane time constant. Each neuron maintains small history buffers for previous membrane potential and prediction-current state.

At each timestep, the instantaneous prediction error (xs/τ)(\bm{x} - \bm{s}/\tau) is low-pass filtered into mp\bm{m}_p and fed back as IpredI_{\rm pred}.

3. Modified Membrane Potential Update and Surrogate Gradient Approach

In discrete-time, the membrane dynamics integrate the prediction current as follows:

  1. Total input current: Il[t]=xl1[t]+mpl[t1]\bm{I}^l[t] = \bm{x}^{l-1}[t] + \bm{m}_p^l[t-1]
  2. Membrane update:

ml[t]=(11τl)vl[t1]+1τlIl[t]\bm{m}^l[t] = \Bigl(1-\frac{1}{\tau^l}\Bigr)\bm{v}^l[t-1] + \frac{1}{\tau^l}\,\bm{I}^l[t]

Spike emission: sl[t]=H(ml[t]θl)\bm{s}^l[t] = H(\bm{m}^l[t] - \theta^l), with membrane potential reset

  1. Prediction-current update:

mpl[t]=(1τpl)mpl[t1]+τpl(xl1[t]sl[t]τl)\bm{m}_p^l[t] = (1-\tau_p^l)\,\bm{m}_p^l[t-1] + \tau_p^l\bigl(\bm{x}^{l-1}[t] - \frac{\bm{s}^l[t]}{\tau^l}\bigr)

To enable gradient-based training despite the Heaviside nonlinearity, a smooth surrogate gradient replaces H/m\partial H/\partial m:

sl[t]ml[t]σ ⁣(ml[t]θl)=kσ(k(mθl))[1σ(k(mθl))]\frac{\partial s^l[t]}{\partial m^l[t]} \approx \sigma'\!\bigl(m^l[t]-\theta^l\bigr) = k\,\sigma\bigl(k(m-\theta^l)\bigr)\left[1-\sigma(k(m-\theta^l))\right]

with slope parameter k10k \approx 10.

4. Gradient Flow: Direct and Indirect Pathways

The augmented neuron introduces two new gradient propagation pathways via the prediction-current state mp\bm{m}_p:

Lxl1[t]=Lml[t]ml[t]xl1[t]+Lmpl[t]mpl[t]xl1[t]\frac{\partial L}{\partial \bm{x}^{l-1}[t]} = \frac{\partial L}{\partial \bm{m}^l[t]}\frac{\partial \bm{m}^l[t]}{\partial \bm{x}^{l-1}[t]} + \frac{\partial L}{\partial \bm{m}_p^l[t]}\frac{\partial \bm{m}_p^l[t]}{\partial \bm{x}^{l-1}[t]}

and

Lml[t]=Lsl[t]sl[t]ml[t]+Lml[t+1]ml[t+1]ml[t]+Lmpl[t]mpl[t]sl[t]sl[t]ml[t]\frac{\partial L}{\partial \bm{m}^l[t]} = \frac{\partial L}{\partial \bm{s}^l[t]}\frac{\partial s^l[t]}{\partial m^l[t]} + \frac{\partial L}{\partial \bm{m}^l[t+1]}\frac{\partial \bm{m}^l[t+1]}{\partial \bm{m}^l[t]} + \frac{\partial L}{\partial \bm{m}_p^l[t]}\frac{\partial \bm{m}_p^l[t]}{\partial \bm{s}^l[t]}\frac{\partial s^l[t]}{\partial m^l[t]}

The direct pathway bypasses the near-zero s/m\partial s/\partial m outside spike events. The indirect pathway modulates prediction error correction at each step via mp/s\partial m_p/\partial s. This dual mechanism alleviates vanishing temporal gradients and stabilizes training.

5. Biological Alignment: Dendritic Modulation and Error-Driven Plasticity

The prediction-current update functionally mimics distal dendritic modulation observed in cortical pyramidal cells, where distal dendrites carry subthreshold, top-down signals that bias membrane potential but do not directly trigger firing. The mp\bm{m}_p term acts analogously by shifting membrane potential toward or away from threshold according to the neuron’s expected spiking. Furthermore, the update rule mpmp+τp(xs/τ)\bm{m}_p \leftarrow \bm{m}_p + \tau_p (x - s/\tau) resembles local, calcium-mediated error-filtering driving synaptic plasticity as observed in experimental studies of NMDA spikes and dendritic signaling (Huang et al., 29 Jan 2026).

6. Implementation Specifications

Layer-wise pseudocode for S-LIF neuron integration is presented as follows:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
initialize W^l, τ^l, θ^l, τ_p^l;  # learnable time constants
initialize v^l[0]=0, m_p^l[0]=0;

for t=1 to T:
  # 1) compute feedforward synaptic input
  I_syn = W^l · s^{l-1}[t];
  
  # 2) add self-prediction current
  I_total = I_syn + m_p^l[t-1];
  
  # 3) update membrane potential
  m^l[t] = (1 - 1/τ^l) * v^l[t-1] + (1/τ^l) * I_total;
  s^l[t] = Heaviside(m^l[t] - θ^l);  # forward spike
  v^l[t] = m^l[t] - s^l[t] * (m^l[t] - v_reset);
  
  # 4) update prediction current
  pred_error = I_syn - s^l[t]/τ^l;
  m_p^l[t] = (1 - τ_p^l) * m_p^l[t-1] + τ_p^l * pred_error;
end for

Notable hyperparameters include τl[2,5]\tau^l \in [2,5], τpl\tau_p^l initialized to 0.1–0.2 (and learned), Adam learning rate 10310^{-3}, and surrogate slope k10k \approx 10. Memory overhead consists of a single additional float per neuron for mpm_p. Computationally, the scheme requires two multiplies and one add per neuron per timestep beyond the standard LIF baseline.

7. Empirical Performance Across Domains

Experimental validation demonstrates performance improvement in classification and control tasks, summarized as follows:

Task Domain Key Targets / Benchmarks Performance Gains
Image Classification CIFAR-10, CIFAR-100, ImageNet-100, ImageNet-1k; ConvNet, SEW-ResNet, Spiking-ResNet +0.3–1.0% top-1 accuracy for T=4–8
Sequential Classification Sequential CIFAR-10 (T=32T=32), various neuron types (IF, LIF, PLIF, CLIF) +2–3% test accuracy
Reinforcement Learning MuJoCo (Ant, HalfCheetah, Hopper, Walker2d), TD3, proxy-target & vanilla SNN −3% (ANN baseline) → +0.7% (vanilla), +3.3% (proxy)

Gains are consistent across networks, time steps, and neuron configurations, indicating broad applicability and closing performance gaps between SNN and ANN frameworks (Huang et al., 29 Jan 2026).


The Self-Prediction Enhanced LIF neuron thus introduces a biologically plausible, computationally minimal augmentation to standard spiking neuron models, creating continuous error-corrective pathways that improve gradient flow and task performance. The design is aligned with core neuroscientific principles, is easily implementable with negligible overhead, and substantively advances SNN training and effectiveness across practical domains.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Self-Prediction Enhanced LIF Neuron.