Hybrid QKAN: Quantum-Inspired KANs
- Hybrid QKAN (HQKAN) is a hybrid neural architecture that combines quantum variational activations with classical Kolmogorov–Arnold Networks to achieve exponential expressivity per parameter.
- It utilizes single-qubit DARUAN modules and latent bottlenecking to significantly reduce parameter counts while supporting scalable architectures like Transformers and LSTMs.
- HQKAN offers tractable gradient-based training through analytic gradients and parameter-shift rules, ensuring compatibility with both classical simulations and near-term quantum devices.
Hybrid Quantum-inspired Kolmogorov–Arnold Networks (HQKANs) are a class of hybrid neural architectures that integrate quantum variational activation functions—realized as single-qubit Data Re-Uploading Activation (DARUAN) modules—within classical Kolmogorov–Arnold Networks (KANs). This hybridization aims to combine quantum-inspired spectral expressivity with classical scalability, yielding models that achieve exponential expressivity per parameter, parameter efficiency, and tractable training even at large scale. The HQKAN framework subsumes feedforward, hierarchical, and recurrent architectures, and is implementable on both classical hardware (via analytic simulation) and near-term quantum devices (Jiang et al., 17 Sep 2025, Hsu et al., 4 Dec 2025, Werner et al., 27 Jun 2025).
1. Mathematical Foundation and Motivation
The Kolmogorov–Arnold representation theorem establishes that any continuous multivariate function can be decomposed into a sum of univariate functions. Classical KANs operationalize this by representing layer outputs as
where each is a learnable univariate activation.
Quantum extensions exploit parameterized variational quantum circuits (VQCs) to realize quantum variational activation functions (QVAFs), which exponentially expand the frequency spectrum accessible for function approximation. The single-qubit Data Re-Uploading Activation (DARUAN) circuit takes a scalar and, via repeated parameterized rotations and data re-encoding, outputs the expectation value
for a parameterized unitary . With trainable Fourier frequencies and depths , the spectrum of representable frequencies grows as (with frequencies , ). Thus, for target error , only circuit parameters are required, representing an exponential reduction relative to classical Fourier activations (Jiang et al., 17 Sep 2025).
Naïve QKAN layers embedded directly in MLP-scale architectures face impractically high parameter counts (scaling quadratically in input/output dimension). HQKAN addresses this via dimensionality bottlenecking combined with a quantum-inspired latent core, maintaining exponential expressivity while drastically reducing total parameterization.
2. HQKAN Architectural Design
The core HQKAN module inserts a low-dimensional (latent) bottleneck between two classical linear maps, sandwiching a compact QKAN (QVAF) core. For an input dimension , output , and latent dimension , a single HQKAN layer is:
- Down-projection: ,
- QKAN core: apply QVAF to each via DARUAN:
- Up-projection: ,
Pseudocode:
1 2 3 4 |
z = A @ x # Down-project φ = [QVAF_DARUAN(z_i, θ_Q[i], ω_Q[i]) for i in range(d_latent)] y = B @ φ # Up-project return y |
For hierarchical and sequence models, the HQKAN latent core can be wrapped within encoder–decoder structures (JHCG Net), supporting stacked or modular abstraction. In recurrent settings such as HQKAN-LSTM, each LSTM gate’s affine map is replaced by a sum of QVAF modules, with optional latent bottlenecking for further compression (Hsu et al., 4 Dec 2025).
3. Training, Initialization, and Optimization
HQKAN models are optimized using standard gradient-based routines (e.g., Adam, L-BFGS), with gradients for quantum circuit parameters obtained via the parameter-shift rule: When simulated classically, QVAF operations can leverage closed-form analytic gradients, yielding compatibility with PyTorch-style autodiff (Hsu et al., 4 Dec 2025). The classical maps are initialized via Glorot procedures; quantum circuit angles are initialized near identity (small-angle rotations) to mitigate barren plateaus; QVAF encoding weights may be set geometrically () or trained.
Losses are task-specific—MSE for regression, cross-entropy for classification, perplexity for autoregressive language modeling. HQKANs offer favorable parameter complexity:
- HQKAN core: parameters (per QVAF depth )
- Classical maps: When , total parameters are significantly reduced compared to MLPs of similar output dimension.
4. Theoretical Properties and Guarantees
- Exponential spectrum expansion: Given geometric weights in a depth- QVAF, the accessible frequencies in the activation spectrum fill the integer interval (Jiang et al., 17 Sep 2025).
- Parameter efficiency: For and error target in -norm, a QVAF(DARUAN) with depth suffices, yielding logarithmic scaling in total quantum parameters, exponentially reducing parameter count compared to Fourier-based KANs (Jiang et al., 17 Sep 2025).
- Hierarchical abstraction and interpretability: HQKANs embedded in encoder–decoder stacks admit explicit frequency-resolved decomposition per layer; in LSTM variants, gate subfunctions can be interpreted as tracking distinct frequency bands (Hsu et al., 4 Dec 2025).
- NISQ-compliance and classical simulability: Single-qubit (non-entangling) circuits allow efficient classical emulation, facilitating deployment on classical hardware or as a backend for future NISQ/QPU hardware (Jiang et al., 17 Sep 2025, Hsu et al., 4 Dec 2025).
5. Empirical Results and Benchmarks
Qualitative and quantitative performance of HQKANs is established via a broad set of benchmarks:
| Task/Model | HQKAN | KAN | MLP | QKAN |
|---|---|---|---|---|
| Regression (RMSE, Feynman 66) | Lowest RMSE, ~30% fewer params | Lower RMSE than MLP | – | – |
| Image Classification (CIFAR-10 Top-1, Top-5) | 71.6%, 97.9% (14,370 params) | – | 68.4%, 97.4% (41,802 params) | 68.8%, 97.0% (21,280 params) |
| Language Modeling (GPT-2 Ppl@41M params) | Lower perplexity, faster convergence | – | Ppl@124M params, same training time | – |
| Sequential (LSTM Cell, Damped SHM MAE/MSE/R2) | Test MSE 4.32e-4 / R2 0.9903 | 1.33e-3 / 0.9701 | – | 1.02e-3 / 0.9771 |
These results consistently show HQKANs either outperforming or matching conventional baselines while utilizing significantly lower parameter counts. Scalability is empirically demonstrated: HQKAN-based models (e.g., KANsformer with Flash-attention) maintain memory usage and iteration times comparable to, or better than, MLP-type architectures when scaling to large batch sizes and distributed hardware (Jiang et al., 17 Sep 2025).
6. Extensions: Hierarchical, Recurrent, and Quantum-Classical Hybrids
- Hierarchical HQKAN/JHCG Net: Implementation of HQKAN as the latent core of encoder–decoder (autoencoder) structures yields stacked abstraction and improved interpretability. Each latent block acts as a KAN with QVAF activations, enabling smooth, additive decompositions that decompose the feature spectrum (Hsu et al., 4 Dec 2025).
- HQKAN-LSTM: Embedding HQKAN gates inside recurrent LSTM structures replaces affine transformations in input, forget, output, and update gates by KAN-style sums of QVAF modules. This achieves lower parameter counts and better R², particularly in regimes dominated by nonlinear temporal dependencies (Hsu et al., 4 Dec 2025).
- Variants (QuKAN, EVQKAN): Related frameworks include QuKAN (Werner et al., 27 Jun 2025), which utilizes quantum circuit Born machines for spline-based KAN residuals within classical networks, and EVQKAN (Wakaura et al., 28 Mar 2025), which constructs variational quantum networks with layerwise tiling for neuromorphic quantum applications.
7. Interpretability, Scalability, and Hardware Deployment
HQKANs are explicitly designed to maintain the interpretability advantage of KANs: each QVAF activation corresponds to a smooth Fourier-enriched subfunction whose spectral content can be analyzed. Additive decomposition permits frequency tracking per layer, and parameter mapping between quantum amplitudes and classical spline coefficients (Werner et al., 27 Jun 2025, Hsu et al., 4 Dec 2025).
Scalability is inherent—non-entangling single-qubit circuits have negligible simulation overhead and are easily vectorized on GPUs. All operations are compatible with classical autodiff frameworks, and swapping in hardware quantum backends (for true quantum estimation) is straightforward at the single-qubit layer level (Jiang et al., 17 Sep 2025, Hsu et al., 4 Dec 2025).
Overall, HQKANs unify the Kolmogorov–Arnold functional decomposition, quantum-inspired variational spectral enrichment, and modern autoencoding hierarchies into a modular and efficient family of architectures. They achieve superior expressivity-per-parameter, reduction in memory and computational requirements, while offering interpretability and extensibility to sequential, hierarchical, and large-scale settings (Jiang et al., 17 Sep 2025, Hsu et al., 4 Dec 2025, Werner et al., 27 Jun 2025).