Expected Star Discrepancy Theory
- Expected star discrepancy theory is the study of the average maximal deviation between empirical point distributions and the uniform measure over anchored, axis-aligned boxes in the unit cube.
- It establishes sharp bounds using i.i.d. sampling with a Θ(√(d/N)) rate and advanced stratified techniques, such as jittered and convex equivolume partitions, achieving improved O(N^(–1/2–1/(2d))) rates.
- The theory underpins practical numerical integration methods, informing quasi-Monte Carlo techniques and optimal partitioning strategies for controlling uniform error in high-dimensional settings.
Expected star discrepancy theory quantifies the typical deviation between the empirical measure of randomly sampled or stratified point sets in the unit cube and the uniform (Lebesgue) measure, as measured by the supremum over axis-aligned anchored boxes. This field rigorously analyses the average case behavior for major sampling techniques, providing sharp bounds and dissecting partition strategies that minimize expected uniform error in high-dimensional numerical integration, such as quasi-Monte Carlo.
1. Formalism and Classical Results
Let . For any , define the anchored box and let denote its volume. The star discrepancy is
Expected star discrepancy refers to , where the average is taken over the randomness inherent in the point set , either due to sampling or stratification.
For independent uniform samples (standard Monte Carlo), it is established that
with lower and upper bounds matching up to constants, as proven via chaining, VC theory, and concentration inequalities (Doerr, 2012, Gómez et al., 2020, Löbbe, 2014). Aistleitner–Hofer provide explicit tail bounds for i.i.d. samples, and the existence of point sets with discrepancy at most is guaranteed by Heinrich–Novak–Wasilkowski–Woźniakowski. However, deterministic constructions matching this remain elusive.
2. Jittered and Stratified Sampling: Theory and Bounds
Jittered sampling partitions into axis-aligned cubes, placing one random point per cube. Pausinger–Steinerberger proved that for constants ,
strictly outperforming the scaling of i.i.d. random sampling appreciably when (Pausinger et al., 2015, Doerr, 2021, Xian et al., 2023). Recent refinements improved constants and precise logarithmic corrections, e.g.,
demonstrating tight control over expected error for practical regimes (Xu et al., 25 Dec 2025). Bernstein-type inequalities, slicing, and chaining arguments are standard analytical tools.
3. The Strong Partition Principle
The strong partition principle asserts that for any partition of into equal-volume cells, stratification yields strictly smaller expected star discrepancy than Monte Carlo: Xian–Xu, Xu–Xian, and others generalize this result beyond axis-aligned partitions—any equal-volume measurable partition delivers variance dominance, and thus expectation dominance, over i.i.d. sampling (Xian et al., 2022, Xian et al., 2023, Xu et al., 25 Dec 2025). The proof leverages variance decomposition for each test box and systemic reduction across boundary effects, and is robust to finite and arbitrary cell geometries. This principle resolves previously open questions in discrepancy theory.
4. Optimal Partitioning: Convex Equivolume Schemes
Advances in partition design have produced explicit classes of convex equivolume partitions (e.g., Stolarsky partitions, ) that outperform classical jittered sampling: where denotes jittered/gridded, i.i.d., and new partition sampling. The optimal expected discrepancy under such tilings obeys
with nontrivial improvements in the constant, sharpening estimates even for moderate , (Xu et al., 29 Dec 2025, Xian et al., 2022). These constructions answer conjectures on the existence of partitions beating jittered sampling in expectation with explicit geometric recipes, and suggest generalizations to even more sophisticated tilings and boundary-shape optimizations.
5. Practical Construction and Applications
Secure pseudorandom bit generators (PRNGs), such as AES–CTR–DRBG, are shown to yield point sets with discrepancy indistinguishable from truly random sets, provided cryptographic security parameter is high and rounding errors controlled (Gómez et al., 2020). Lacunary point sets, generated by recurrences like , achieve comparable rates (with a penalty), but enable fast deterministic generation from a single random seed, advantageous in hardware or reproducibility-critical contexts (Löbbe, 2014).
Quasi-Monte Carlo integration benefits directly from stratification: lower expected star discrepancy translates into smaller uniform error for indicator test functions, justifying widespread adoption of stratified and randomized-QMC schemes in high-dimensional numerical integration.
6. Lower Bounds, Tightness, and Open Problems
Doerr (Doerr, 2012) establishes the matching lower bound for expected star discrepancy of random i.i.d. samples: for some absolute , confirming the law is tight. For stratified/jittered constructions, the scaling and the role of partition geometry in lowering constants is provably sharp. Open problems concern explicit deterministic constructions matching best random bounds, optimization of leading constants for new partition schemes, and extension of the strong partition principle to alternative notions of discrepancy (e.g., weighted or directional variants) (Xian et al., 2022, Xu et al., 25 Dec 2025, Xu et al., 29 Dec 2025).
7. Summary Table: Expected Star Discrepancy Rates
| Sampling Method | Expected Rate | Partition Principle Validity |
|---|---|---|
| Simple random (Monte Carlo, i.i.d.) | Dominated by stratified schemes | |
| Jittered sampling (axis-aligned cubes) | , constants improved recently | Strictly better than MC |
| Convex equivolume partitions | , strictly smaller constant | Strong partition principle holds |
| Secure PRNG-based sets (high-security) | (almost indistinguishable from random) | Follows MC behavior |
| Lacunary sequences (e.g., mod 1) | Slight penalty over MC |
These rates define the frontier of expected star discrepancy theory, which now centers on realizing sharp constants, advanced partitioning, and extending the partition principle to broader families of sampling schemes and discrepancy types. Stratification—not just axis-aligned but convex or tailored—is universally preferable to pure random sampling for minimizing expected uniformity error in finite dimensions.