Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
GPT-5.1
GPT-5.1 104 tok/s
Gemini 3.0 Pro 36 tok/s Pro
Gemini 2.5 Flash 133 tok/s Pro
Kimi K2 216 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

VeriBToT: Robust Boosting Verification

Updated 19 November 2025
  • VeriBToT is a framework for training, certifying, and verifying gradient-boosted tree classifiers under adversarial perturbations with exact robustness guarantees.
  • It leverages the large-spread ensemble property to enable efficient, linear or pseudo-polynomial time certification against norm-bounded adversarial attacks.
  • By imposing structural constraints during training, VeriBToT maintains predictive accuracy with less than 1% loss, while significantly accelerating verification compared to conventional methods.

VeriBToT is a framework for training, certifying, and verifying gradient-boosted tree classifiers under adversarial perturbations, enabling exact robustness guarantees in polynomial or pseudo-polynomial time under specific conditions. Central to VeriBToT is the large-spread ensemble property, which enables efficient certification against norm-bounded adversarial attacks for advanced tree boosting models, including those trained with XGBoost or LightGBM. By imposing structural constraints during training and deploying tailored verification algorithms, VeriBToT addresses the computational complexity of robustness certification without significant degradation to predictive accuracy (Calzavara et al., 22 Feb 2024).

1. Formal Model and Definitions

VeriBToT targets binary classification over a feature space XRdX \subseteq \mathbb{R}^d and label set Y={+1,1}Y = \{+1,-1\}. The fundamental model is a gradient-boosted ensemble T={t1,,tm}T = \{t_1, \dots, t_m\}, where each tit_i is a regression tree. Each regression tree tt is recursively defined as either a leaf node λ(s)\lambda(s) with real-valued score ss, or an internal node σ(f,v,tL,tR)\sigma(f,v,t_L,t_R) splitting at feature ff and threshold vv. The overall ensemble prediction on input xx is:

T^(x)=i=1mti(x)\hat T(x) = \sum_{i=1}^m t_i(x)

Classification applies a monotone link ι:RR\iota: \mathbb{R} \to \mathbb{R} and threshold τ\tau, returning T(x)=+1T(x) = +1 if ι(T^(x))τ\iota(\hat T(x)) \geq \tau and 1-1 otherwise.

The defining property for enabled verification is "large-spread": for norm p\lVert\cdot\rVert_p, TT is large-spread with respect to adversarial budget kk if

spreadp(T)>2k\text{spread}_p(T) > 2k

where the pp-spread quantifies the minimal separation (in the pp-norm) between any pair of thresholds on the same feature across distinct trees.

2. Complexity and Verifiability Results

Verification of classifier robustness under adversarial attack consists of deciding, for instance xx and ground-truth yy, whether every allowed perturbation zz with zxpk\lVert z - x \rVert_p \leq k leaves classification invariant (T(z)=yT(z) = y). For large-spread ensembles:

  • For the \ell_\infty-norm, exact verification is achieved in linear time (O(N)O(N), NN = total number of tree nodes). This stems from the lack of interference between trees under large-spread orthogonality.
  • For any fixed p<p < \infty, robustness verification is NP-hard. Nonetheless, pseudo-polynomial time algorithms exist using dynamic programming.

These properties extend verifiable learning principles to boosted ensembles.

3. Verification Algorithms

a. \ell_\infty-Norm: Linear-Time Verification

The attacker's optimal strategy consists of, for each tree, selecting the highest-gain reachable leaf (with perturbation cost k\leq k). For all tit_i:

  • Identify leaves jj with minimal perturbation δijk\lVert \delta_{ij}\rVert_\infty \leq k.
  • Compute gain GijG_{ij} in ensemble score.
  • Aggregate maximal non-negative GijG_{ij} across trees to form Γ\Gamma:

Γ=i=1mmaxjLimax(0,Gij)\Gamma = \sum_{i=1}^m \max_{j \in L_i} \max(0, G_{ij})

Then, for raw score s=T^(x)s = \hat T(x), robustness is certified if:

  • y=+1y=+1: ι(sΓ)τ\iota(s - \Gamma) \geq \tau,
  • y=1y=-1: ι(s+Γ)<τ\iota(s + \Gamma) < \tau.

The algorithm is linear in the number of nodes:

1
2
3
4
5
6
7
8
9
function VERIFY_LINF(T, x, y, k):
  if T(x) != y: return False
  Γ = 0
  for each tree t_i in T:
    compute reachable leaves L_i and gains G_{ij}
    Γ += max_{j  L_i} max(0, G_{ij})
  s = sum_i t_i(x)
  if y == +1: return (i(s - Γ) >= τ)
  else:       return (i(s + Γ) < τ)

b. General p\ell_p-Norm: Pseudo-Polynomial Verification

With p\ell_p, the adversarial budget is iwiK\sum_i w_i \leq K, where wi=δippw_i = \lVert \delta_i \rVert_p^p and K=kpK = k^p. Each tree presents leaf/gain pairs (wij,Gij)(w_{ij}, G_{ij}). This yields a grouped-knapsack DP:

M[i,q]=max{M[i1,q], maxj:wijq(M[i1,qwij]+Gij)}M[i,q] = \max\Bigl\{ M[i-1, q],\ \max_{j: w_{ij}\le q}(M[i-1, q-w_{ij}]+G_{ij}) \Bigr\}

Initialize M[0,q]=0M[0,q]=0 for all q=0,,Kq=0,\dots,K. The answer is Γ=M[m,K]\Gamma=M[m,K]; the same classification test applies as for \ell_\infty.

The complexity is O(mKLmax)O(m \cdot K \cdot L_{\text{max}}), where LmaxL_{\text{max}} is the number of leaves per tree. Weight discretization is used when wijw_{ij} and KK are not integers.

4. Practical Construction of Large-Spread Ensembles

During training—e.g., in LightGBM—whenever a tree tit_i splits on ff at threshold vv, subsequent trees are barred from choosing v(v2k,v+2k)v' \in (v-2k, v+2k) for feature ff. This ensures the large-spread condition throughout the ensemble. Such exclusions marginally reduce model capacity, but empirical results indicate that predictive accuracy loss remains negligible (within 1%1\% of standard boosting).

5. Empirical Performance and Limitations

VeriBToT-certified models have been benchmarked on FMNIST, MNIST, and Webspam. Key findings:

  • Accuracy of large-spread models matches that of unconstrained LightGBM within 1%1\%.
  • Robustness certified by VeriBToT matches or nearly matches the most optimistic bounds of unconstrained GBDT, but is guaranteed for every test point.
  • Verification time is accelerated by 10×10\times to 400×400\times relative to MILP or abstract-interpretation approaches, even for ensembles with up to 125 trees (each of depth 8).
  • Computation for 0\ell_0 or 1\ell_1 attacks is pseudo-polynomial in kpk^p, but this is practical for small perturbation budgets.

6. Significance and Conclusions

VeriBToT extends verifiable machine learning from majority-vote tree ensembles to more powerful boosted ensemble techniques, retaining provable robustness against norm-bounded adversaries in feasible computational time. Imposing the large-spread constraint during training does not materially compromise accuracy, while enabling efficient and exact certification under \ell_\infty and, with dynamic programming, for general p\ell_p attacks. This makes VeriBToT suitable for high-assurance applications requiring both model performance and certified robustness (Calzavara et al., 22 Feb 2024).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)
Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to VeriBToT.