Hadamard-Domain Convolution
- Hadamard-domain convolution is a transform-based operation that uses the ±1 Hadamard transform to convert convolution into elementwise multiplication, offering speed and energy efficiency.
- It reduces computational complexity to O(n log n) by diagonalizing convolution operations, making it ideal for low-power hardware and scalable deep neural network implementations.
- Recent studies demonstrate its versatility in enhancing neural network layers, model compression, time series feature extraction, and even fractional calculus computations.
Hadamard-domain convolution refers to a class of convolution-like operations performed using the Hadamard (or Walsh–Hadamard) transform or, in a distributional analysis context, to multiplicative convolution (in contrast to the classical additive convolution). This approach leverages the algebraic and computational properties of the Hadamard transform—an orthogonal linear transform with elements ±1—which enables highly efficient, multiplication-free, and in some cases non-parametric implementations of convolutional and pointwise operators in a variety of domains, including signal processing, deep neural networks, time series analysis, and fractional calculus. Both the foundational aspects and practical algorithmic constructions of Hadamard-domain convolution are rich and technically diverse.
1. Mathematical Foundations and Definitions
The Hadamard transform, denoted as for size-, is a real orthogonal matrix with entries in . It satisfies , and is constructed recursively. The Walsh–Hadamard transform (WHT) of a vector is , and the inverse is .
There are two major senses of "convolution" in the Hadamard domain:
- Multiplicative (Hadamard) Convolution on Temperate Distributions: For (the space of temperate distributions), the *-product is defined by duality as
where . This operation generalizes classical convolution to multiplicative group actions (Vogt, 2018).
- Transform-Domain Convolution (Signal and Neural Networks): The Hadamard (Walsh) transform diagonalizes the dyadic (Walsh) or circulant convolution:
- Dyadic convolution: For ,
with being bitwise XOR. By the Hadamard convolution theorem,
where denotes elementwise multiplication (Pan et al., 2023). - Pointwise and 2D Convolutional Analogues: By transforming inputs and kernels to the Hadamard domain, convolution becomes elementwise multiplication, analogous to the Fourier convolution theorem but using real-valued basis (Mannam, 2022, Pan et al., 2022, Pan et al., 2021, Jeong et al., 2019).
2. Hadamard-Convolution Operators on Distribution Spaces
Hadamard-type operators are linear maps such that every monomial is an eigenvector. Vogt's Theorem establishes that is of Hadamard type if and only if it is convolution (in the multiplicative sense above) with a unique , the space of distributions with "θ–rapid decay" both at infinity and near the coordinate hyperplanes (Vogt, 2018). Explicitly, this means is a finite sum of derivatives of functions with
for all multi-indices and integers . The exponential transform maps the multiplicative convolution structure on the positive quadrant to the usual additive convolution on a Fréchet space of exponentially decreasing functions, revealing deep connections between these operator classes and classical translation-invariant convolution (Vogt, 2018).
3. Hadamard-Domain Convolution in Neural Networks
Hadamard-domain convolutional layers arise in several neural network constructions:
- 1D/2D Hadamard Layers: Input tensors are transformed along the channel or spatial axes via FWHT, nonlinearity (e.g., smooth-thresholding or soft-thresholding), then inverse FWHT (Pan et al., 2021, Pan et al., 2022).
- Hadamard-Perceptron/HT-Block: In 2D, is applied across rows and columns, followed by trainable scaling, channel mixing (1×1 convolutions), and soft-thresholding in the transform domain, then inverse (Pan et al., 2023).
- DWHT Pointwise Convolution: Replaces 1×1 learned convolution by a fixed DWHT of the channel vector, with zero learnable parameters and only additions per spatial location (Jeong et al., 2019).
- Block WHT for Arbitrary Channel Sizes: Applies 1D WHTs to blocks of channels with overlapping windows, avoiding wasteful zero padding, and is especially useful when channel counts are not powers of two (Pan et al., 2022).
- Comparison to Fourier-Domain Convolution: Analogous to FFT-based acceleration, spatial convolution becomes trivial elementwise multiplication in the transform domain. For Hadamard, the kernels and activations remain real, and the transform involves no multiplications, yielding significant speed and parameter advantages (Crasmaru, 2018).
These approaches are particularly effective when:
- Hardware requires low latency and energy (e.g., IoT, edge devices) (Mannam, 2022).
- Model compression and runtime are prioritized over marginal accuracy losses.
- Multiplication-free operations are essential, as on low-cost DSP, FPGA, or ASIC platforms (Pan et al., 2021, Jeong et al., 2019).
4. Computational Properties and Empirical Performance
Hadamard-domain convolution leverages the following computational features:
| Convolution Type | Complexity per Location | Parameters | Multiplies |
|---|---|---|---|
| Standard 1×1 conv | |||
| Hadamard/WHT-based 1D | or $0$ | $0$ (adds only) | |
| 2D WHT-based | $0$ (adds only) | ||
| FFT-based conv |
- Speed and Memory: For typical MobileNet or ResNet blocks, time per forward pass with WHT layer is up to faster and uses up to less RAM (batch ), e.g., $1.886$ s for conv vs.\ $0.077$ s for 2D WHT (Pan et al., 2022).
- Parameter Reduction: Replacing 1/3 of 1×1 convs in MobileNet-V2 yields a reduction in trainable parameters with only $1.75$ percentage points accuracy loss (CIFAR-10) (Pan et al., 2022).
- Energy Consumption: For kernel size , energy per pass for Hadamard method is lower than direct convolution when not too large, due to vastly reduced multiplies—only the elementwise product in the transform domain requires multiplications (Mannam, 2022).
- Empirical Results: On MNIST, Hadamard-domain and classical convolution achieve near-identical test accuracy; on CIFAR-10/100 with multi-channel images, the Hadamard method underperforms by 2–9 points depending on architecture and layer replacement ratio (Mannam, 2022, Pan et al., 2021, Jeong et al., 2019).
5. Variant Constructions and Practical Implementations
Several practical variants and algorithmic improvements have been developed:
- Smooth-Thresholding and Denoising: Nonlinearities such as are applied in the transform domain, often with per-coefficient (but not per-connection) learned thresholds (Pan et al., 2021, Pan et al., 2022). These serve as efficient, channel-wise regularizers.
- Multiplication-Free Depthwise Convolution: Local block operations, such as and related operators, realize depthwise convolution using only sign and addition operations (Pan et al., 2021).
- Learned Permutations and Generalized Bases: One can replace the fixed Hadamard basis by learned orthogonal matrices for data-dependent transforms (Mannam, 2022).
- Quantum Implementation: The Hadamard transform is realizable by Hadamard gates in quantum circuits; hybrid quantum-classical schemes thereby perform some or all of the convolutional transform steps on a quantum processor (Pan et al., 2023).
- Time Series Feature Extraction: Columns of Hadamard matrices are used as orthogonal, ±1-valued convolution kernels in the ROCKET feature-extraction framework, yielding state-of-the-art noise-robustness and at least 2× faster training compared to random-kernel approaches (Hao et al., 3 Nov 2025).
6. Extensions, Limitations, and Theoretical Implications
Extensions:
- Fractional Calculus: Hadamard convolution appears in Caputo–Hadamard fractional derivative discretization. Convolution quadrature methods for subdiffusion equations are extended to the logarithmic (Hadamard) case, with high-order correction for initial singularities (Yin et al., 2023).
- Full Frequency-Domain Networks: Expressing all layers as transform-domain operations (e.g., via DFT or DWT) using DFT/Hadamard convolution theorems allows backpropagation entirely in the frequency domain, with specialized activations and loss (Crasmaru, 2018).
Limitations:
- Accuracy-Complexity Tradeoff: On high-dimensional, multi-channel, or highly textured inputs, Hadamard-based methods exhibit increased error relative to learned spatial convolution unless more sophisticated transforms or hybrid approaches are employed (Mannam, 2022, Pan et al., 2021).
- Power-of-Two Constraints: Direct (F)WHT requires padding of non-power-of-two channel/spatial sizes, potentially introducing inefficiency or output truncation (Pan et al., 2022).
- Factored vs. Locality: Walsh (dyadic) or circulant convolution differs structurally from spatial () convolution, affecting feature localization—HT-blocks only approximate true spatial relationships (Pan et al., 2023).
- Backward Compatibility and Layer Cascade: Cascading multiple Hadamard-domain layers often requires repeated forward/inverse transforms, unless staying exclusively in the transform domain (Mannam, 2022, Crasmaru, 2018). A plausible implication is that end-to-end Hadamard networks may benefit from trainable or hybrid frequency-domain operations.
Theoretical context: In the distributional setting, Hadamard-domain convolution provides a complete classification of all continuous linear operators on for which monomials are eigenvectors, revealing connections to both classical analysis and functional-analytic structure (Vogt, 2018).
7. Summary and Significance
Hadamard-domain convolution unifies a broad spectrum of transform-based operations across mathematical analysis, signal processing, neural network architecture, and numerical methods for differential equations. Its defining property—the transformation of convolution or correlation to elementwise operations in the Hadamard basis—enables orders-of-magnitude reductions in memory, energy, and computational complexity, especially in resource-constrained environments. Recent works illustrate both rigorous mathematical classification for operator theory (Vogt, 2018) and diverse algorithmic proposals for deep learning (Pan et al., 2022, Pan et al., 2021, Jeong et al., 2019), low-power vision (Mannam, 2022), quantum-classical hybrid systems (Pan et al., 2023), and robust, efficient time-series feature engineering (Hao et al., 3 Nov 2025). The breadth of methods and empirical successes underline the Hadamard domain's enduring centrality in efficient and theoretically grounded convolutional computation.