Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 91 tok/s
Gemini 2.5 Pro 58 tok/s Pro
GPT-5 Medium 29 tok/s
GPT-5 High 29 tok/s Pro
GPT-4o 102 tok/s
GPT OSS 120B 462 tok/s Pro
Kimi K2 181 tok/s Pro
2000 character limit reached

Uniformity Tests via Overlapping Spacings

Updated 31 August 2025
  • The paper establishes a rigorous asymptotic framework for uniformity tests based on sum-functions of overlapping spacings, extending classical methods to multidimensional settings.
  • It demonstrates that the test performance is governed by the correlation between overlapping and disjoint spacings, with explicit Pitman efficacy formulas quantifying efficiency.
  • The study shows that increasing the spacing order improves discrimination power and reduces variance, moving detection rates closer to the parametric n⁻¹/² limit under local alternatives.

Tests for uniformity based on sum-functions of overlapping spacings form a central class of nonparametric test statistics, extending the classic spacings philosophy from univariate to multidimensional, discrete, and circular settings. These tests have recently been the focus of advanced asymptotic studies, particularly when the order of overlapping spacings grows with the sample size, and their efficacy strongly depends on connections with disjoint spacings statistics. Applications range from classical goodness-of-fit in one dimension to uniformity testing on spheres, high-dimensional data, and spatial point processes.

1. Statistical Framework for Higher Order Overlapping Spacings

The principle underlying tests based on sum-functions of overlapping spacings is to use the ordered sample (X1,...,Xn)(X_1, ..., X_n) (typically after applying a probability integral transform to reduce the problem to [0,1][0,1]) and consider mm-order overlapping spacings: Dk,m=Xk+mXk,0knmD_{k,m} = X_{k+m} - X_k, \quad 0 \leq k \leq n-m where mm may increase with nn. The prototypical test statistic takes the form

Vn,m=k=0n1h(nDk,m)V_{n,m} = \sum_{k=0}^{n-1} h(n D_{k,m})

with h()h(\cdot) a chosen function (e.g., h(x)=x2h(x)=x^2 for the Greenwood statistic) (Mirakhmedov, 26 Aug 2025). Overlapping spacings differ from disjoint spacings by their dependence structure: overlapping spacings share sample points, leading to more data reuse and stronger correlation between terms.

A corresponding "disjoint spacings" version partitions the data into non-overlapping blocks: Vn,m=k=0Nh(nDkm,m)V_{n,m}^* = \sum_{k=0}^{N} h(n D_{k\cdot m,m}) where N=n/m1N = n/m - 1, and Dkm,mD_{k\cdot m,m} are mm-step disjoint spacings.

Overlapping spacings allow the order mm to diverge with nn, typically m=o(n)m = o(n), to achieve improved discrimination between uniformity and alternatives.

2. Asymptotic Distribution and Local Power

Under the null hypothesis of uniformity, statistics Vn,mV_{n,m} can be shown to be asymptotically normal when properly centered and normalized (Mirakhmedov, 26 Aug 2025, Mirakhmedov, 15 Apr 2024): Vn,mnA0,nσmndN(0,1)\frac{V_{n,m} - n\mathcal{A}_{0,n}}{\sigma_m\sqrt{n}} \xrightarrow{d} N(0,1) where A0,n=E[h(Z0,m)]\mathcal{A}_{0,n} = E[h(Z_{0,m})] (with Z0,mZ_{0,m} being the sum of mm standard exponentials) and σm2\sigma_m^2 depends on the variance and covariances of h(Z0,m)h(Z_{0,m}) with its nearby shifts, as well as a correction for the mean-variance relationship created by overlap: σm2=Var[h(Z0,m)]+2j=1m1Cov(h(Z0,m),h(Zj,m))m2τm2\sigma_m^2 = \mathrm{Var}[h(Z_{0,m})] + 2\sum_{j=1}^{m-1}\mathrm{Cov}(h(Z_{0,m}), h(Z_{j,m})) - m^2\tau_m^2 with τm=m1Cov(h(Z0,m),Z0,m)\tau_m = m^{-1}\mathrm{Cov}(h(Z_{0,m}), Z_{0,m}) (Mirakhmedov, 26 Aug 2025, Mirakhmedov, 15 Apr 2024).

Under local alternatives of the form f(x)=1+(nm)1/4ln(x)f(x) = 1 + (nm)^{-1/4} l_n(x), the expectation of Vn,mV_{n,m} is shifted by

A1,n=E[h(Z0,m)]+σmm+12nμm(h)l22\mathcal{A}_{1,n} = E[h(Z_{0,m})] + \frac{\sigma_m^*\sqrt{m+1}}{\sqrt{2n}} \mu_m(h) \lVert l \rVert_2^2

where σm2=Var[φ(Z0,m)]\sigma_m^{*2} = \mathrm{Var}[\varphi(Z_{0,m})] for φ(u)=h(u)E[h(Z0,m)](um)τm\varphi(u) = h(u) - E[h(Z_{0,m})] - (u-m)\tau_m, and the efficacy is quantified by

em2(h)=(m+1)σm22σm2μm2(h)e_m^2(h) = \frac{(m+1) \sigma_m^{*2}}{2\sigma_m^2} \mu_m^2(h)

with

μm(h)=corr(φ(Z0,m),(Z0,mm)2)\mu_m(h) = \mathrm{corr}(\varphi(Z_{0,m}), (Z_{0,m} - m)^2)

The power is then given by

PowerΦ(em(h)l22uα)\text{Power} \approx \Phi\left(e_m(h)\lVert l \rVert_2^2 - u_\alpha\right)

with uα=Φ1(1α)u_\alpha = \Phi^{-1}(1-\alpha).

The key parameter μm(h)\mu_m(h)—correlation between the centered function of the spacing and its quadratic deviation—drives the local asymptotic power (Mirakhmedov, 26 Aug 2025).

3. Comparison and Role of Disjoint Spacings Statistics

A critical insight is that the asymptotic power of overlapping spacings tests is governed by the efficacy of the statistics based on disjoint spacings. Under similar Lyapunov-type conditions, disjoint spacings statistics Vn,mV_{n,m}^* satisfy

E[Vn,m]=NA0,n(1+o(1)),Var(Vn,m)=Nσm2(1+o(1))E[V_{n,m}^*] = N\mathcal{A}_{0,n}(1+o(1)), \qquad \mathrm{Var}(V_{n,m}^*) = N\sigma_m^{*2}(1+o(1))

with efficacy under local alternatives

em2(h)=m+12mμm2(h)e_m^{*2}(h) = \frac{m+1}{2m}\mu_m^2(h)

The Pitman relative efficiency between overlapping and disjoint spacings test is

PE(Vn,m(h),Vn,m(h))=limnmσm2(h)σm2(h)1PE(V_{n,m}(h), V_{n,m}^*(h)) = \lim_{n\to\infty} \frac{m \sigma_m^{*2}(h)}{\sigma_m^2(h)} \geq 1

Thus, overlapping spacings tests are at least as efficient (typically more so): for the Greenwood statistic, the disjoint test requires approximately $1.5$ times the sample size to reach equivalent power (Mirakhmedov, 26 Aug 2025). This dependence reveals overlapping spacings tests can be calibrated and understood via their disjoint analogues.

4. Effect of Increasing Spacing Order

Permitting the order mm of spacings to increase with nn (subject to m=o(n)m = o(n)) significantly impacts the rate at which local alternatives can be discriminated. With fixed mm, alternatives detectable at n1/4n^{-1/4} are at the statistical limit, but if mm diverges, the rate improves to (nm)1/4(nm)^{-1/4}, approaching the parametric n1/2n^{-1/2} rate in the limit. This improvement is apparent in the efficacy expressions, for instance: $e_m^2(h) \to \frac{3}{4} \text{ for the Greenwood statistic (as %%%%36%%%%)}$ Hence, higher-order overlapping spacings “average over” larger sample intervals, reducing variance and enhancing the test's ability to detect local deviations from uniformity.

A plausible implication is that practitioners can tune mm to balance power and stability, achieving higher sensitivity for alternatives with finer-grained deviations as mm increases—subject to computational constraints and the requirement m=o(n)m=o(n).

5. Practical Relevance and Pitman Efficiency

These findings provide detailed guidance for designing and analyzing uniformity tests based on sum-functions of overlapping spacings. The core facts are:

  • Overlapping spacings tests admit a normal limit for a wide class of functions hh and for orders m=o(n)m=o(n) (Mirakhmedov, 15 Apr 2024, Mirakhmedov, 26 Aug 2025).
  • Their asymptotic local power depends strongly on correlation structures rooted in disjoint spacings statistics, directly quantifiable and interpretable via μm(h)\mu_m(h) and Pitman ARE formulas.
  • The choice of hh (Greenwood for quadratic, Moran for log spacings, etc.) determines the direction of maximal efficacy, and standard choices yield locally most powerful tests within the flexible class of overlapping spacings statistics (Singh et al., 2021).
  • For fixed mm, efficiency is limited, but exercises in the recent literature show that increasing mm sharpens discrimination rates and heightens Pitman efficacy, up to the $3/4$ bound for Greenwood-type statistics (Mirakhmedov, 26 Aug 2025).
  • The asymptotic calibration is exact enough to support principled sample size and power calculations.

This suggests that, for nonparametric goodness-of-fit problems where high sensitivity to fine-scale deviations from uniformity is required, overlapping spacings methods with suitably chosen or diverging orders are theoretically preferred.

6. Mathematical Formulas and Summary Table

The principal formulas used in the asymptotic theory are summarized below:

Statistic Formula under Null and Alternatives Efficacy / Power
Overlapping Vn,mV_{n,m} Vn,mnA0,nσmnN(0,1)\displaystyle \frac{V_{n,m} - n\mathcal{A}_{0,n}}{\sigma_m\sqrt{n}} \to N(0,1) em2(h)=(m+1)σm22σm2μm2(h)\displaystyle e_m^2(h) = \frac{(m+1)\sigma_m^{*2}}{2\sigma_m^2} \mu_m^2(h)
Disjoint Vn,mV_{n,m}^* Same as above, N=n/m1N = n/m - 1, σm2\sigma_m^{*2} replaces σm2\sigma_m^2 em2(h)=m+12mμm2(h)\displaystyle e_m^{*2}(h) = \frac{m+1}{2m} \mu_m^2(h)

Key parameters:

  • A0,n=E[h(Z0,m)]\mathcal{A}_{0,n} = E[h(Z_{0,m})]
  • σm2\sigma_m^2 includes variance and covariance terms of h(Z0,m)h(Z_{0,m})
  • μm(h)\mu_m(h): correlation of φ(Z0,m)\varphi(Z_{0,m}) with quadratic deviations
  • Pitman ARE: PE(Vn,m(h),Vn,m(h))=mσm2(h)/σm2(h)PE(V_{n,m}(h), V_{n,m}^*(h)) = m \sigma_m^{*2}(h)/\sigma_m^2(h)

These precise expressions should be used to calculate sample size, select the order mm, and choose the tuning function hh for optimal uniformity testing.

7. Conclusion

Tests for uniformity based on sum-functions of overlapping spacings, especially those allowing the spacing order to grow with sample size, are theoretically robust and highly efficient. The recent asymptotic theory establishes their normal limiting distribution under mild conditions and clarifies the dependence of their power and efficiency on related disjoint spacings statistics. The framework accommodates flexible choices of hh and supports tuning of mm for improved discrimination power. The critical link to disjoint spacings statistics, the explicit Pitman efficacy formulas, and the normal limiting distribution position these methods as a theoretically optimal choice for modern nonparametric testing of uniformity in a variety of applied statistical domains (Mirakhmedov, 26 Aug 2025, Mirakhmedov, 15 Apr 2024).