ℓ0-Ball Convex Hull for Robust Neural Verification
- The topic defines the ℓ0-ball as a nonconvex set formed by a union of coordinate-flats and describes its convex hull via an intersection with a scaled ℓ1-polytope.
- It details how the convex hull is obtained by intersecting the axis-aligned bounding box with an asymmetrically scaled ℓ1-like polytope, achieving tighter relaxations than traditional methods.
- The analysis introduces a computationally efficient top‑k strategy that significantly improves neural network robustness verification against sparse adversarial attacks.
The convex hull of an -ball is a central object in the formal verification of neural network robustness against few-pixel (sparse) adversarial attacks. Unlike the convex and well-understood -balls for , the -ball comprises a finite union of -dimensional flats in and is highly nonconvex for . Recent work establishes that the convex hull of an -ball can be described precisely as the intersection of its axis-aligned bounding box and an asymmetrically scaled -like polytope, enabling tight geometric and computational characterizations that outperform previous relaxations in both accuracy and tractability (Shapira et al., 13 Nov 2025).
1. Definition and Nonconvexity of the -Ball
In , the centered -ball of radius about a reference point is
where denotes the number of nonzero coordinates in . Geometrically, consists of all points differing from in at most coordinates, forming a union of all -dimensional axis-aligned subspaces ("coordinate-flats"). For , this set is highly nonconvex and discrete in structure, posing challenges in the application of standard convex relaxation techniques used in neural network robustness certification.
2. Convex Hull Characterization via Intersection
To analyze and exploit for verification, it is necessary to work with its convex hull. Let each coordinate be constrained within and define the ambient box . The convex hull is shown to satisfy:
where is a "scaled -polytope," given by
with
Each measures, asymmetrically, how far the th coordinate moves away from , normalized over its permissible interval. The intersection excludes those points of the box requiring more than coordinates to be displaced maximally, thus tightly bounding the convex hull of the sparse attack set.
3. Geometric Properties and Volume Analysis
A direct orthant-decomposition gives closed-form volume formulas:
- .
- .
The ratio of excess volume between and tends to zero exponentially as (fixed ), demonstrating that is a geometrically tight superset of for high-dimensional input spaces. In contrast, the bounding box includes a super-polynomial excess in volume as dimension increases, leading to significant looseness when used for adversarial budget relaxation.
4. Exact Linear Bound Propagation over
Propagation of affine bounds through neural networks typically relies on overapproximating the perturbation domain. For any linear form over : Let
The minimum and maximum optimizer is given by:
where and denote the th smallest/largest entries among and , respectively. Thus, the propagation procedure simply selects the top- coordinates with the most extreme contributions, matching the combinatorial nature of itself.
5. Algorithmic Integration and Comparative Analysis
In practical neural network verification settings, these "top-" updates are integrated into standard linear-bound propagation frameworks (e.g., GPUPoly), replacing the sign-checks used for or domains. All other GPU-friendly sum-and-reduce kernels remain intact except for parallel tracking of the extremal values.
Comparison of relaxation techniques for bounding yields:
| Relaxation | Bound Method | Tightness/Empirical Performance |
|---|---|---|
| Box-only | Sum all values | Gross over-approximation, adversary over-limited |
| Pure | Multiply largest by | Looser than actual top-, poorly modeled limits |
| Conv() | Sum top- extremal or | Tight domain, empirically 3x–7x more properties proven |
On MNIST, Fashion-MNIST, and CIFAR-10 benchmarks, the top- propagation method proves – as many local-robustness properties within the same time budget.
6. Geometric and Computational Advantages
is strictly smaller than both the bounding box and the -relaxation, deviating on only a negligible fraction of the box's total volume for typical high-dimensional inputs. Algorithmically, optimal linear-programming over requires only or time by top- selection, matching the efficiency of existing sum-reduce primitives and avoiding the combinatorial explosion associated with the -flat union in itself.
7. Implications and Application in Robustness Pipelines
The precise geometric and computational properties of enable a robust certified-robustness pipeline for attacks on deep networks that is both tighter and computationally faster than prior box- or -based relaxations. In practice, this improvement leads to substantially tighter neuron-range estimates in bound-propagation designs, significantly accelerating verification designs (including CAV/CAVerification stages) without scalability loss.
A plausible implication is the generalization of these characterizations to broader sparse perturbation models or other structured nonconvex sets, as the intersection-of-box-and-polytope principle offers a template for tight approximations within convex formal verification frameworks.
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days free