Papers
Topics
Authors
Recent
2000 character limit reached

Discretized NeuTucF Model for Sparse Turbulence Data

Updated 12 December 2025
  • The paper introduces a neural variant of Tucker factorization that discretizes continuous wind field data, achieving state-of-the-art performance in turbulence estimation.
  • It employs quantile-based binning and a four-mode neural Tucker decomposition to embed discrete indices and capture complex spatio-temporal interactions.
  • Empirical results show lower MAE and RMSE compared to baselines, demonstrating robust imputation and regression of atmospheric turbulence quantities.

The discretized NeuTucF model is a neural variant of Tucker factorization tailored to model sparse, multi-dimensional turbulence data using quantized, discrete input features. Developed to estimate turbulence quantities such as the Richardson number from wind profile radar data, NeuTucF achieves state-of-the-art performance on continuous yet sparse three-dimensional wind fields by embedding discretized indices into low-rank representations and capturing multiway spatio-temporal interactions via a four-mode neural Tucker decomposition. This methodology enables imputation and regression of missing entries in real-world atmospheric datasets, outperforming common baseline models (Fan et al., 5 Dec 2025).

1. Input Discretization and Quantization

The model operates on four continuous variables: altitude (hh), and three wind speed components (uu, vv, ww). Each variable undergoes the following transformation pipeline:

  • Standardization: Each feature XX is z-normalized,

Xstd=Xμσ,X_{\rm std} = \frac{X-\mu}{\sigma},

where μ\mu and σ\sigma are the empirical mean and standard deviation.

  • Quantile-based binning: Each standardized feature is partitioned into KK equally populated bins. Quantile boundaries BkB_k are computed as

Bk=Q(kK),k=0,1,,K,B_k = Q\left(\frac{k}{K}\right),\quad k=0,1,\dots,K,

where Q()Q(\cdot) denotes the empirical quantile function.

  • Piecewise discretization: Each real-valued standardized input xx is mapped to its corresponding bin index using

idx(x)={0,x<B1, i,Bix<Bi+1,i=1,,K1, K,xBK.\mathrm{idx}(x)= \begin{cases} 0, & x < B_1, \ i, & B_i \le x < B_{i+1},\quad i=1,\dots,K-1, \ K, & x\ge B_K. \end{cases}

Applying this scheme to all features produces discrete indices p,i,j,k{0,,K}p, i, j, k \in \{0,\dots,K\} for each observation. All quantized data are then aggregated into a four-way array

Z=D(Y)R(K+1)×(K+1)×(K+1)×(K+1),\mathcal Z = \mathcal D(\mathcal Y) \in \mathbb R^{(K+1)\times (K+1)\times (K+1)\times (K+1)},

with Zp,i,j,k\mathcal Z_{p,i,j,k} encoding the (possibly missing) Richardson number at the joint quantized coordinates.

2. Tucker-Based Neural Model Architecture

The NeuTucF model encodes interactions between quantized features using a four-mode neural Tucker factorization:

  • Factor matrices (embeddings): For each mode n=1,2,3,4n=1,2,3,4 (corresponding to h,u,v,wh,u,v,w), a learnable matrix

U(n)R(K+1)×rU^{(n)} \in \mathbb R^{(K+1) \times r}

provides rr-dimensional embeddings for each discrete value. For instance, e(p)=Up,:(1)e^{(p)} = U^{(1)}_{p,:} embeds bin index pp of height.

  • Core tensor: A dense core tensor

GRr×r×r×r\mathcal G \in \mathbb R^{r \times r \times r \times r}

encodes higher-order interactions between modes. All experiments fix r=5r=5.

  • Tucker reconstruction: Classically, the estimated value at (p,i,j,k)(p,i,j,k) is

y^pijk=a,b,c,dGabcdUp,a(1)Ui,b(2)Uj,c(3)Uk,d(4).\hat y_{pijk} = \sum_{a,b,c,d} G_{abcd} U^{(1)}_{p,a} U^{(2)}_{i,b} U^{(3)}_{j,c} U^{(4)}_{k,d}.

3. Neural Interaction Tensor Construction

Rather than multiplying full mode-factor matrices as in classical Tucker, NeuTucF constructs the "interaction tensor" using outer products of the relevant embeddings: Tpijk=e(p)e(i)e(j)e(k)Rr×r×r×r,\mathcal T_{pijk} = e^{(p)} \circ e^{(i)} \circ e^{(j)} \circ e^{(k)} \in \mathbb R^{r\times r\times r\times r}, where \circ denotes the outer product. Entrywise,

Tpijk[a,b,c,d]=ea(p)  eb(i)  ec(j)  ed(k).\mathcal T_{pijk}[a,b,c,d] = e^{(p)}_a\; e^{(i)}_b\; e^{(j)}_c\; e^{(k)}_d.

The flattened interaction tensor vec(Tpijk)\mathrm{vec}(\mathcal T_{pijk}) is then linearly projected using the flattened core vec(G)\mathrm{vec}(\mathcal G): y^pijk=σ(vec(G)vec(Tpijk)),\hat y_{pijk} = \sigma \left( \mathrm{vec}(\mathcal G)^\top \mathrm{vec}\bigl(\mathcal T_{pijk}\bigr) \right), where σ\sigma is a sigmoid or identity activation. This neural mapping allows fully expressive four-way nonlinear interactions parameterized by the core tensor.

4. Training Objective and Optimization

Training minimizes the mean squared error (MSE) over all observed entries Ω\Omega: L=1Ω(p,i,j,k)Ω(y^pijkypijk)2+λ(GF2+n=14U(n)F2),\mathcal L = \frac{1}{|\Omega|} \sum_{(p,i,j,k)\in\Omega} \left( \hat y_{pijk}-y_{pijk} \right)^2 + \lambda \left( \|\mathcal G\|_F^2 +\sum_{n=1}^4\|U^{(n)}\|_F^2 \right), where ypijky_{pijk} are ground-truth Richardson numbers and λ\lambda controls L2L_2 weight regularization.

Implementation utilizes PyTorch and the Adam optimizer (learning rate 1e ⁣ ⁣31\mathrm e\!-\!3), training for approximately 100 epochs in minibatches. This configuration suits high-dimensional, sparse regression over wind field tensors.

5. Inference and Reconstruction

After training, the model can impute or regress missing entries (p,i,j,k)Ω(p,i,j,k)\notin\Omega by passing their respective discrete indices through the same embedding, outer product, and linear mapping pipeline. The inference formula is equivalent to Tucker-factor reconstruction: y^pijk=a,b,c,dGabcdUp,a(1)Ui,b(2)Uj,c(3)Uk,d(4).\hat y_{pijk} = \sum_{a,b,c,d} G_{abcd} U^{(1)}_{p,a} U^{(2)}_{i,b} U^{(3)}_{j,c} U^{(4)}_{k,d}. This unified parametrization allows estimation anywhere within the quantized input domain without explicit tensor completion.

6. Empirical Results and Performance Metrics

Model efficacy is evaluated via five-fold cross-validation using mean absolute error (MAE), root-mean-square error (RMSE), and coefficient of determination (R2R^2):

Model MAE (\downarrow) RMSE (\downarrow) R2R^2 (\uparrow)
M1 (NeuTucF) 11.64±2.4711.64 \pm 2.47 20.11±3.7620.11 \pm 3.76 0.353±0.1950.353 \pm 0.195
Baselines (M2–M5) Higher Higher Lower

All baseline models (M2–M5) exhibit higher MAE and RMSE and lower R2R^2 compared to NeuTucF under matched embedding size and Tucker rank (r=5r=5). The chosen discretization and low-rank parametric structure yield the best cross-validated error; no explicit ablation of bin count or Tucker rank is reported, but these hyperparameters are fixed for fair comparison (Fan et al., 5 Dec 2025).

7. Context, Significance, and Limitations

The discretized NeuTucF approach addresses the challenge of estimating turbulence quantities from sparse, incomplete, or irregularly-sampled atmospheric measurements, particularly in settings where only wind profile radar is available. Discretization adapts continuous inputs for compatibility with embedding-based neural tensor factorization models, while the full four-way Tucker core captures complex spatio-temporal dependencies. The method demonstrates robust imputational and regression accuracy across a variety of low-altitude turbulence datasets. A plausible implication is that similar discretized neural factorization architectures could be adapted for regression tasks involving other high-dimensional, spatio-temporal geophysical data.

However, the model's effectiveness is contingent on the appropriateness of the chosen binning strategy and Tucker rank. The paper reports no explicit sensitivity analysis on these hyperparameters, leaving open the question of how choices such as KK (the number of quantile bins) affect generalization or bias (Fan et al., 5 Dec 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Whiteboard

Follow Topic

Get notified by email when new papers are published related to Discretized NeuTucF Model.