Papers
Topics
Authors
Recent
2000 character limit reached

RocketStack: Deep Recursive Ensemble

Updated 6 December 2025
  • RocketStack is a deep recursive ensemble learning framework that extends stacking architectures with up to ten layers, integrating predictions through adaptive model pruning and feature compression.
  • The framework leverages recursive stacking, noise-perturbed pruning, and periodic feature compression methods (such as SFE, autoencoder, and attention-based selection) to control computational and feature complexity.
  • Empirical results on binary and multi-class datasets show significant accuracy improvements (up to 6.11% gains) and substantial runtime and feature dimensionality reductions compared to traditional stacking methods.

RocketStack is a level-aware deep recursive ensemble learning framework designed to extend the depth of stacking architectures while controlling computational and feature complexity through adaptive model pruning, feature compression, and stochastic regularization. The methodology systematically advances beyond conventional horizontal diversity in ensemble learning by enabling recursive stacking up to ten levels, thus promoting deeper representational integration across base learners with tractable computational costs (Demirel, 20 Jun 2025).

1. Recursive Stacking Architecture

RocketStack generalizes traditional stacking by constructing a hierarchy of ensemble layers, each integrating predictions from the preceding level through meta-feature concatenation and selective pruning. Let X(0)Rn×dX^{(0)} \in \mathbb{R}^{n \times d} represent the original nn-sample, dd-feature training set, with Xte(0)X_{\rm te}^{(0)} as its hold-out counterpart. At stacking level \ell, the ensemble consists of M(1)={m1,...,mM1}\mathcal{M}^{(\ell-1)} = \{ m_1, ..., m_{M_{\ell-1}} \}, where M1M_{\ell-1} is the number of models retained post-pruning from the previous level.

Each model undergoes KK-fold cross-validation; its concatenated out-of-fold (OOF) predictions form pi()Rnp_i^{(\ell)} \in \mathbb{R}^n. Aggregating predictions across all models yields P()Rn×M1P^{(\ell)} \in \mathbb{R}^{n \times M_{\ell-1}}. The iterative meta-feature expansion is defined as:

X()=[X(1)P()],Xte()=[Xte(1)m(Xte(1))]X^{(\ell)} = [X^{(\ell-1)} \mid P^{(\ell)}],\quad X^{(\ell)}_{\rm te} = [X^{(\ell-1)}_{\rm te} \mid m(X^{(\ell-1)}_{\rm te})]

where \mid denotes column-concatenation and m()m(\cdot) the vector of hold-out predictions.

Model pruning is performed at each level to ensure M<M1M_\ell < M_{\ell-1}. Raw OOF performance scores ai()a_i^{(\ell)} (accuracy or AUC) are computed for each mim_i, and a custom threshold Qcustom()Q^{(\ell)}_{\rm custom} is defined as the quantile at γ=5+80[std(a~())]2\gamma = 5 + 80[\mathrm{std}(\tilde{\mathbf a}^{(\ell)})]^2, where a~i()\tilde a_i^{(\ell)} may be the raw or noise-perturbed score. Only models meeting a~i()Qcustom()\tilde a_i^{(\ell)} \ge Q^{(\ell)}_{\rm custom} are retained.

2. Pruning Strategies and Feature Compression Mechanisms

A key innovation in RocketStack is the introduction of noise-perturbed pruning and adaptive feature compression, implemented as follows:

Noise-perturbed Pruning

Mild Gaussian noise is added to OOF scores prior to pruning to serve as a regularizer:

a~i()=ai()+ϵi,ϵiN(0,λrange(a()))\tilde a_i^{(\ell)} = a_i^{(\ell)} + \epsilon_i,\quad \epsilon_i \sim \mathcal{N}(0,\,\lambda \cdot \mathrm{range}(\mathbf a^{(\ell)}))

where λ{0,0.05,0.10}\lambda \in \{0, 0.05, 0.10\}. Strict (λ=0\lambda=0) and randomized (λ>0\lambda>0) schemes are compared.

Feature Compression

Feature dimensionality is controlled either at every level or periodically (e.g., levels 3, 6, 9), using one of three compressors:

  • Simple, Fast, Efficient (SFE) Filter: Utility U(f)=Rel(f)/(1+Red(f))\mathcal{U}(f) = \mathrm{Rel}(f)/(1+\mathrm{Red}(f)); select features with highest utility.
  • Autoencoder (AE) Compression: Nonlinear reduction using fθ,gϕ:f_\theta, g_\phi: minimize XX^2\|X - \hat X\|^2 with bottleneck kd/3k \approx d/3.
  • Attention-Based Selection: Compute α=softmax(WX+b)\alpha = \mathrm{softmax}(W X + b); keep xix_i with αiQ75(α)\alpha_i \ge Q_{75}(\alpha).

A simplified pseudocode of the framework orchestrates the OOF generation, optional feature compression, model evaluation, noise injection, and dynamic pruning per level, with user-specified settings for stacking depth LL, cross-validation folds KK, pruning noise λ\lambda, compression mode, compressor type, periodicity, and minimum model count.

3. Model Training, Meta-Learner Pooling, and Computational Complexity

At each level, retained base learners are re-trained on the augmented feature matrix X()X^{(\ell)}, recursively constructing deeper meta-representations. Rather than a single fixed meta-learner, the ensemble at each level comprises all surviving models M()\mathcal{M}^{(\ell)}, with optional selection of the top-kk or the singular top performer for inference.

The computational complexity of each level is dominated by O(KM1Ttrain(d1))O(K M_{\ell-1} T_{\text{train}}(d_{\ell-1})) for cross-validated training, O(d1n)O(d_{\ell-1} n) for filter-based compression (or O(d1nh)O(d_{\ell-1} n h) for autoencoders), and O(M1logM1)O(M_{\ell-1} \log M_{\ell-1}) for pruning. Sublinear runtime growth with increasing \ell is achieved through aggressive pruning and feature reduction, supporting practical exploration to depths of =10\ell = 10.

4. Empirical Evaluation across Binary and Multi-Class Datasets

Experiments across 33 OpenML datasets (23 binary, 10 multi-class) demonstrate the efficacy and scalability of RocketStack:

Binary Classification (Periodic SFE at Levels 3/6/9)

  • Strict pruning (λ=0\lambda=0): 88.08% accuracy at level 10
  • Light randomization (λ=0.05\lambda=0.05): 88.40% (+0.32%)
  • Runtime reduction: \sim10.5% compared to no compression
  • Feature count at L10: \sim6 (vs. 177 with no compression)

Multi-Class Classification (Periodic Attention)

  • Strict pruning: 93.29%
  • Light randomization: 93.67% (+0.38%)
  • Ultimate accuracy at L10: 98.60% (vs. 92.49% best baseline; +6.11%)
  • Runtime reduction: -56.1% relative to no compression
  • Feature reduction at L10: From \sim145 to \sim38 (-74%)

Linear mixed model analysis indicates significant accuracy increases with stacking depth in most configurations (p<.001p < .001). Periodic compression schemes yield the strongest trends (p<.01p < .01), while per-level compression often lacks a significant trend (p>.05p > .05).

5. Staged Ensemble Dynamics: The Rocket Analogy

RocketStack is conceptualized around the metaphor of multistage rocket engineering, encapsulated as “Prune – Compress – Propel”:

  • Prune: Analogous to jettisoning empty fuel tanks, underperforming learners are removed to prevent superfluous complexity.
  • Compress: Periodic feature compression parallels stage separation, allowing informative meta-features to accumulate before redundancy is discarded.
  • Propel: Mild Gaussian randomization in pruning induces a controlled instability, analogous to guidance feedback in rocket dynamics, promoting diversity and mitigating the risk of premature convergence.

These coordinated mechanisms facilitate deep recursive ensembling with sustainable complexity, enabling superior predictive performance relative to shallower and horizontally-diverse stacking architectures.

6. Significance and Implementation Considerations

RocketStack establishes a scalable paradigm for deep ensemble integration, demonstrating that controlled regularization and staged dimensionality reduction can overcome saturation and complexity barriers that previously limited the depth of stack-based learning. Its modular design accommodates advances in feature compression, meta-learner architectures, and adaptive pruning for continued empirical and theoretical exploration (Demirel, 20 Jun 2025). The detailed pseudocode and equation definitions provided in the original manuscript enable rigorous reimplementation and comparative benchmarking.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)
Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to RocketStack.