Papers
Topics
Authors
Recent
2000 character limit reached

High-Dimensional Convolution

Updated 2 December 2025
  • High-dimensional convolution is an extension of classical convolution that generalizes the operator to D dimensions, enabling efficient function approximation and structured kernel designs.
  • It mitigates exponential growth in computation and memory by leveraging sparse strategies and hybrid kernel patterns, reducing FLOPs and enhancing scalability.
  • This framework underpins advanced neural network architectures and non-Euclidean models, improving performance in image processing, 3D recognition, and discrete optimization.

High-dimensional convolution generalizes classical convolution operations beyond the usual one, two, or three spatial dimensions, thereby encompassing a range of mathematical, computational, and algorithmic frameworks with extensive applications in machine learning, numerical analysis, optimization, and geometry. The following sections detail major paradigms, theoretical properties, practical instantiations, and computational implications as developed in recent arXiv literature.

1. Mathematical Foundations and Operator Generalizations

High-dimensional convolution on ZD\mathbb{Z}^D is defined as an extension of the classical linear operator. Let DD denote ambient dimension, x:ZDRCinx: \mathbb{Z}^D \rightarrow \mathbb{R}^{C_{\mathrm{in}}} the input, WiRCout×CinW_i \in \mathbb{R}^{C_{\mathrm{out}} \times C_{\mathrm{in}}} a collection of kernel weights indexed by offset iNDZDi \in \mathcal{N}^D \subset \mathbb{Z}^D (kernel support), and Cin,Cout\mathcal{C}^{\mathrm{in}}, \mathcal{C}^{\mathrm{out}} sparse (or dense) sets of input and output integer coordinates. The generalized convolution is

xuout=iND:u+iCinWixu+iin,uCoutx^{\mathrm{out}}_u = \sum_{i \in \mathcal{N}^D: u+i \in \mathcal{C}^{\mathrm{in}}} W_i\, x^{\mathrm{in}}_{u+i}, \quad \forall u \in \mathcal{C}^{\mathrm{out}}

This structure admits dense and sparse, strided or dilated, hyper-cubic or arbitrary kernel supports. In high-dimensional function approximation, a convolution kernel Ψc,h(x)\Psi_{\mathbf{c},\mathbf{h}}(\mathbf{x}) is often factorized as a tensor product of univariate kernels over each axis: Ψc,h(x)=k=1dΨck,hk(xk)\Psi_{\mathbf{c},\mathbf{h}}(\mathbf{x}) = \prod_{k=1}^d \Psi_{c_k, h_k}(x_k) with smoothness and localization ensured through parameter choice, enabling efficient regularization and approximation in high-dimensional domains (Choy et al., 2019, Gao et al., 22 Sep 2024).

2. Computational Complexity and Sparse Implementation Strategies

A principal computational challenge for high-dimensional convolution is the exponential growth of tensor sizes, kernel supports, and memory requirements:

  • Dense computation: For a side length SS, kernel size KK, and DD dimensions, the cost is O(SDKDCinCout)O(S^D K^D C_{\mathrm{in}} C_{\mathrm{out}}) FLOPs, O(SD)O(S^D) memory.
  • Sparse strategy: If only NN input voxels are populated (e.g., 3D/4D LIDAR), coordinate-feature pairs are quantized and stored as (C,F)(C, F), with CC the list of N×DN \times D coordinates and FF the N×CinN \times C_{\mathrm{in}} features. The per-voxel cost is then O(NNDCinCout)O(N|\mathcal{N}^D|C_{\mathrm{in}}C_{\mathrm{out}}) (Choy et al., 2019).

Kernel designs such as the "hybrid" pattern in 4D, employing a small 3D spatial block and a 1D temporal cross,

Nhyb4={(i,j,k,0):i,j,krs}{(0,0,0,τ):τrt}\mathcal{N}^4_{\mathrm{hyb}} = \{ (i,j,k,0): |i|,|j|,|k|\le r_s \} \cup \{ (0,0,0,\tau): |\tau|\le r_t \}

achieve substantial parameter and FLOP reductions, empirical speedups (1.3x over full 4D convolutions for r=1r=1), and improved generalization (Choy et al., 2019).

3. High-Dimensional and Higher-Order Convolutional NN Architectures

Standard high-dimensional convolutions are used in convolutional neural network (CNN) backbones for spatial-temporal and volumetric data:

  • The Minkowski Engine supports 1D to 4D convnets with full autograd via sparse-tensor BLAS, batched hash lookups, and efficient kernel-mapping (Choy et al., 2019).
  • In density functional neural approximations, fully convolutional architectures implement mappings from densities ρ(r)\rho(\mathbf{r}) to correlation functions c1(r)c_1(\mathbf{r}) in 2D and 3D, with convolutions as weighted densities and learnable back-projections (Glitsch et al., 19 Feb 2025).

Higher-order convolutions generalize linear convolution to Volterra–like expansions, incorporating polynomials of local patch features: y(p)=b+p=1dpWp,xppy(p) = b + \sum_{p=1}^d \nabla_p \langle W^p, x_p^{\otimes p} \rangle For each order pp, WpW^p is an order-pp tensor and xppx_p^{\otimes p} the pp-fold patch outer product. This architecture efficiently captures higher-order image statistics—e.g., 3rd- and 4th-order correlations important in natural image processing—while achieving optimal performance at modest parameter counts (Azeglio et al., 9 Dec 2024).

4. Algebraic and Non-Euclidean High-Dimensional Convolutions

Vector-map convolutions (V-MapConv):

  • Generalize complex (2D) and quaternion (4D) convolutions to arbitrary D by weight-sharing and learnable inter-subspace mixing. The parameterization reduces from O(D2CinCoutK2)O(D^2 C_{in} C_{out} K^2) to O(DCinCoutK2+D2)O(D C_{in} C_{out} K^2 + D^2), maintains entanglement via a D×DD \times D mixing matrix, and is empirically superior in parameter economy and test performance for large-D data (e.g., satellite bands) (Gaudet et al., 2020).

Volumetric convolution in non-Euclidean domains:

  • On the unit ball B3\mathbb{B}^3, convolution must respect SO(3)SO(3) symmetry. This is accomplished by projecting functions onto a Zernike polynomial basis, performing convolution via coefficient multiplication and rotation-invariant re-projection: (fg)(α,β)=4π3n,,mΩnm(f)Ωn0(g)Ym(α,β)(f * g)(\alpha, \beta) = \frac{4\pi}{3} \sum_{n,\ell,m} \Omega_{n\ell m}(f) \Omega_{n\ell 0}(g) Y_{\ell m}(\alpha, \beta) This construction is differentiable, highly structured, efficiently implemented, and enables explicit symmetry measurement, offering advantages for shape analysis and 3D recognition (Ramasinghe et al., 2019).

5. High-Dimensional Convolution in Approximation Theory and Numerical Analysis

In high-dimensional function approximation, convolution serves as a regularization primitive for quasi-interpolants:

  • A two-step procedure convolves the function with a localized, smooth kernel (controlling "convolution error"), then discretizes with a sparse-grid quadrature ("discretization error"): QN[f](x)=i=1NwiΨc,h(xxi)f(xi)Q_N[f](\mathbf{x}) = \sum_{i=1}^N w_i \Psi_{\mathbf{c},\mathbf{h}}(\mathbf{x}-\mathbf{x}_i) f(\mathbf{x}_i) For a Smolyak sparse grid, the number of nodes grows as O(h1(lnh1)d1)O(h^{-1} (\ln h^{-1})^{d-1}), mitigating the curse of dimensionality compared to tensor grids (O(hd)O(h^{-d})). Achieved rates are O(Ns(lnN)s(d1)+1)O(N^{-s} (\ln N)^{s(d-1)+1}), and empirical results demonstrate near-predicted convergence even for d10d \geq 10 (Gao et al., 22 Sep 2024).

6. High-Dimensional Convolution for Discrete Optimization and Parameterized Algorithms

The (max,+)(\max,+) convolution extends naturally to dd dimensions for applications in multi-constraint discrete optimization: Cv=max0uv(Au+Bvu),v,uNdC_v = \max_{0 \leq u \leq v} \big(A_u + B_{v-u}\big), \quad v, u \in \mathbb{N}^d This high-dimensional formulation is polynomially equivalent (for fixed dd) to multi-constraint $0$-$1$ and unbounded Knapsack and related decision problems. For fixed dimension, parameterized algorithms achieve

O(d(n+Dmax{Π(t+1),tmaxlogtmax}))O \Big(d\, (n + D\,\max \{ \Pi(t+1), t_{\max}\log t_{\max} \}) \Big)

where DD is the number of distinct weights and Π(t+1)=i=1d(ti+1)\Pi(t+1) = \prod_{i=1}^d (t_i+1). The equivalence cycle between convolution, Knapsack, and associated decision problems also yields conditional quadratic lower bounds (assuming the 1D (max,+)(\max,+)-convolution hypothesis) for high-dimensional convolution algorithms (Grage et al., 24 Mar 2024).

7. Empirical Results and Practical Considerations

Empirical evidence underscores the benefits and limitations of high-dimensional and higher-order convolution:

  • On large-scale sparse 3D and 4D datasets (e.g., ScanNet, Synthia), high-dimensional sparse convnets (e.g., MinkowskiNet42, MinkNet20/32) outperform 2D–3D hybrid or pointwise alternatives, with significantly better mIoU and runtime (Choy et al., 2019).
  • Higher-order convolutions (HoCNN, HoResNet-18) provide consistent gains (e.g., 88.13%89.30%88.13\%\to 89.30\% accuracy in image classification) and exhibit qualitatively richer representations as validated via representational similarity analysis (Azeglio et al., 9 Dec 2024).
  • In function approximation, sparse-grid convolution-based quasi-interpolants yield rapid algebraic convergence for dd up to $10$, with strong robustness and efficiency (Gao et al., 22 Sep 2024).
  • Neural DFT models using pure convolutional architectures achieve near-simulation accuracy for density predictions in 2D and generalize efficiently to 3D, speedup factors up to 20×20\times over naive local-ML computations, with the main bottleneck being consistent training data generation in high DD (Glitsch et al., 19 Feb 2025).

8. Outlook and Ongoing Challenges

Ongoing challenges in high-dimensional convolution include:

  • Managing exponential scaling in kernel size, memory, and computation as dimension increases.
  • Designing kernel parameterizations (e.g., hybrid patterns, multi-scale or attention-based kernels) to efficiently capture relevant geometric or statistical dependencies without incurring full KDK^D costs.
  • Developing symmetry-aware architectures for non-Cartesian domains (e.g., balls, spheres, SE(3)), including further efficient implementations of steerable or equivariant kernels.
  • Addressing complexities in optimization problems (e.g., multi-constraint Knapsack, ILPs) where hardness equivalence with high-dimensional convolution persists.
  • Enhancing training and data preparation pipeline scalability for scientific and engineering applications requiring high-dimensional structured data processing.

High-dimensional convolution unifies several active research frontiers and is central to contemporary advances in geometry-aware deep learning, large-scale function approximation, discrete optimization, and physics-informed machine learning (Choy et al., 2019, Gao et al., 22 Sep 2024, Azeglio et al., 9 Dec 2024, Glitsch et al., 19 Feb 2025, Grage et al., 24 Mar 2024, Gaudet et al., 2020, Ramasinghe et al., 2019).

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to High-Dimensional Convolution.