VeriBToT: Robust Boosting Verification
- VeriBToT is a framework for training, certifying, and verifying gradient-boosted tree classifiers under adversarial perturbations with exact robustness guarantees.
- It leverages the large-spread ensemble property to enable efficient, linear or pseudo-polynomial time certification against norm-bounded adversarial attacks.
- By imposing structural constraints during training, VeriBToT maintains predictive accuracy with less than 1% loss, while significantly accelerating verification compared to conventional methods.
VeriBToT is a framework for training, certifying, and verifying gradient-boosted tree classifiers under adversarial perturbations, enabling exact robustness guarantees in polynomial or pseudo-polynomial time under specific conditions. Central to VeriBToT is the large-spread ensemble property, which enables efficient certification against norm-bounded adversarial attacks for advanced tree boosting models, including those trained with XGBoost or LightGBM. By imposing structural constraints during training and deploying tailored verification algorithms, VeriBToT addresses the computational complexity of robustness certification without significant degradation to predictive accuracy (Calzavara et al., 22 Feb 2024).
1. Formal Model and Definitions
VeriBToT targets binary classification over a feature space and label set . The fundamental model is a gradient-boosted ensemble , where each is a regression tree. Each regression tree is recursively defined as either a leaf node with real-valued score , or an internal node splitting at feature and threshold . The overall ensemble prediction on input is:
Classification applies a monotone link and threshold , returning if and otherwise.
The defining property for enabled verification is "large-spread": for norm , is large-spread with respect to adversarial budget if
where the -spread quantifies the minimal separation (in the -norm) between any pair of thresholds on the same feature across distinct trees.
2. Complexity and Verifiability Results
Verification of classifier robustness under adversarial attack consists of deciding, for instance and ground-truth , whether every allowed perturbation with leaves classification invariant (). For large-spread ensembles:
- For the -norm, exact verification is achieved in linear time (, = total number of tree nodes). This stems from the lack of interference between trees under large-spread orthogonality.
- For any fixed , robustness verification is NP-hard. Nonetheless, pseudo-polynomial time algorithms exist using dynamic programming.
These properties extend verifiable learning principles to boosted ensembles.
3. Verification Algorithms
a. -Norm: Linear-Time Verification
The attacker's optimal strategy consists of, for each tree, selecting the highest-gain reachable leaf (with perturbation cost ). For all :
- Identify leaves with minimal perturbation .
- Compute gain in ensemble score.
- Aggregate maximal non-negative across trees to form :
Then, for raw score , robustness is certified if:
- : ,
- : .
The algorithm is linear in the number of nodes:
1 2 3 4 5 6 7 8 9 |
function VERIFY_LINF(T, x, y, k): if T(x) != y: return False Γ = 0 for each tree t_i in T: compute reachable leaves L_i and gains G_{ij} Γ += max_{j ∈ L_i} max(0, G_{ij}) s = sum_i t_i(x) if y == +1: return (i(s - Γ) >= τ) else: return (i(s + Γ) < τ) |
b. General -Norm: Pseudo-Polynomial Verification
With , the adversarial budget is , where and . Each tree presents leaf/gain pairs . This yields a grouped-knapsack DP:
Initialize for all . The answer is ; the same classification test applies as for .
The complexity is , where is the number of leaves per tree. Weight discretization is used when and are not integers.
4. Practical Construction of Large-Spread Ensembles
During training—e.g., in LightGBM—whenever a tree splits on at threshold , subsequent trees are barred from choosing for feature . This ensures the large-spread condition throughout the ensemble. Such exclusions marginally reduce model capacity, but empirical results indicate that predictive accuracy loss remains negligible (within of standard boosting).
5. Empirical Performance and Limitations
VeriBToT-certified models have been benchmarked on FMNIST, MNIST, and Webspam. Key findings:
- Accuracy of large-spread models matches that of unconstrained LightGBM within .
- Robustness certified by VeriBToT matches or nearly matches the most optimistic bounds of unconstrained GBDT, but is guaranteed for every test point.
- Verification time is accelerated by to relative to MILP or abstract-interpretation approaches, even for ensembles with up to 125 trees (each of depth 8).
- Computation for or attacks is pseudo-polynomial in , but this is practical for small perturbation budgets.
6. Significance and Conclusions
VeriBToT extends verifiable machine learning from majority-vote tree ensembles to more powerful boosted ensemble techniques, retaining provable robustness against norm-bounded adversaries in feasible computational time. Imposing the large-spread constraint during training does not materially compromise accuracy, while enabling efficient and exact certification under and, with dynamic programming, for general attacks. This makes VeriBToT suitable for high-assurance applications requiring both model performance and certified robustness (Calzavara et al., 22 Feb 2024).
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days free