Epps–Pulley Statistic: Normality Test
- The Epps–Pulley statistic is a test of normality based on the weighted L²-distance between the empirical characteristic function of standardized data and that of a standard normal distribution.
- It naturally extends to multivariate settings via affine-invariant standardizations and is equivalent to a biased Gaussian-kernel maximum mean discrepancy under simple transformations.
- Its well-understood asymptotic behavior, tunable smoothing parameter β, and spectral representation enable efficient p-value computation and diagnostic use in diverse modeling applications.
The Epps–Pulley statistic is a test of normality constructed as a weighted -distance between the empirical characteristic function of standardized data and that of the standard normal distribution, with a Gaussian weight parameterized by a bandwidth parameter . It extends naturally to multivariate (affine-invariant, universally consistent) settings and is closely related, under simple transformations, to the biased Gaussian-kernel maximum mean discrepancy (MMD) between an empirical distribution and the standard normal. The Epps–Pulley statistic is distribution-free under normality and has well-understood asymptotic behavior, efficiency, and practical performance across diverse alternatives to normality (Ebner et al., 2021, Ebner et al., 2021, Rustamov, 2019).
1. Definition and Forms of the Epps–Pulley Statistic
For a univariate i.i.d. sample with unknown mean and variance , let and . The scaled residuals are . The empirical characteristic function is , and the Gaussian weight is .
The univariate Epps–Pulley statistic is
which can be explicitly computed as
For the multivariate case, if have sample mean and covariance matrix , then standardized residuals are used and the integration occurs over with (Ebner et al., 2021, Ebner et al., 2021). This generalization preserves affine invariance and universal consistency.
2. Connection to Maximum Mean Discrepancy (MMD)
The Epps–Pulley statistic coincides, up to notation and centering, with the biased Gaussian-kernel MMD between an empirical distribution and the standard normal :
with the correspondence (Rustamov, 2019). The Epps–Pulley (or equivalently, the BHEP) statistic thus inherits the omnibus consistency and interpretability of the MMD, where the smoothing parameter (equivalently, kernel width ) modulates sensitivity to different departures from normality.
3. Asymptotic Null Distribution and Eigenstructure
Under , the Epps–Pulley statistic is asymptotically distribution-free. As ,
where is a centered Gaussian process with covariance kernel
(Ebner et al., 2021, Ebner et al., 2021). The limiting law admits the series representation
where are i.i.d. and are the positive eigenvalues of the integral operator on defined by . Numerical values for include (Ebner et al., 2021). This spectral representation underpins tail approximations, p-value computations, and Bahadur slopes.
4. Bahadur Efficiency and Comparative Power
The Bahadur efficiency of the Epps–Pulley statistic against local alternatives indexed by is characterized by two quantities:
- , the asymptotic mean under ,
- , the leading eigenvalue from the null distribution.
Specifically, for small ,
(Ebner et al., 2021). Table 1 in (Ebner et al., 2021) demonstrates that for in the range $0.5 - 1$, the efficiency of the Epps–Pulley test is competitive with or generally superior to EDF-based tests (e.g., Shapiro–Wilk, Anderson–Darling) for certain alternatives, especially those exhibiting skewness (Ley–Paindaveine alternatives):
| β | Lehmann | 1st Ley–P. | 2nd Ley–P. | Contam. N(1,1) | Contam. N(.5,1) | Contam. N(0,.5) |
|---|---|---|---|---|---|---|
| 0.25 | .996 | .947 | .824 | .760 | .945 | .084 |
| 0.5 | .895 | .944 | .872 | .649 | .824 | .267 |
| 0.75 | .854 | .998 | .986 | .592 | .766 | .474 |
| 1 | .743 | .937 | .981 | .499 | .654 | .587 |
Values are efficiencies relative to the (infeasible) likelihood-ratio test (Ebner et al., 2021). This suggests optimal power for near unity across a range of alternatives.
5. Practical Guidelines for Tuning and Implementation
The smoothing parameter modulates the Epps–Pulley test’s sensitivity:
- Small ($0.25$–$0.5$): Emphasizes high-frequency deviations in the characteristic function, suitable for location-type alternatives and mild skew.
- Intermediate ($0.75$–$1$): Effective across both skew and scale-contamination alternatives; aligns with classical BHEP settings.
- Large ($2$–$5$): Favours low-frequency discrepancies, enhancing sensitivity to scale or heavy-tailed alternatives at the cost of skew-detection (Ebner et al., 2021, Rustamov, 2019).
Implementation requires selection or cross-validation of , or adoption of empirically motivated heuristics (e.g., Henze–Zirkler’s for high-dimensional settings). Under the composite null, centering and whitening are performed prior to evaluation (Rustamov, 2019).
Monte Carlo can be used to estimate finite-sample thresholds or compute the Bahadur efficiency via plug-in estimates of and . In practice, standardized versions using explicit null mean/variance formulas allow rapid computation and interpretation (Rustamov, 2019).
6. Comparative and Theoretical Properties
The Epps–Pulley statistic is affine-invariant and, in the multivariate case, universally consistent against alternatives with finite second moments (Ebner et al., 2021, Rustamov, 2019). Compared to Shapiro–Wilk and Anderson–Darling, its key distinctions are:
- Direct applicability in arbitrary dimensions,
- Closed-form critical value calculations via eigen-expansions,
- Explicit power modulated by the smoothing parameter,
- One-to-one equivalence (up to bias) with the widely used MMD, situating it within reproducing kernel Hilbert space (RKHS) theoretical frameworks.
Under the simple null , standardized Epps–Pulley values are approximately ; under the composite null, centering/whitening preserves distribution-freeness up to small sample effects (Rustamov, 2019).
A plausible implication is that the flexibility of the smoothing parameter allows the Epps–Pulley statistic to target classes of alternatives of interest, further justifying its use in modern model criticism and generative modeling settings.
7. Computational Considerations and Extensions
The computation of the Epps–Pulley statistic is explicit and involves only matrix operations and exponential functions. For limiting distribution evaluation, the eigenvalues can be computed numerically by truncating the spectral expansion, with software implementation benefiting from the work of Ebner and Henze (Ebner et al., 2021). For p-values, Imhof’s method or cumulant-matching to the Pearson system is effective.
In the context of generative models such as Wasserstein Auto-Encoders, batch-averaged and standardized forms of the Epps–Pulley (MMD) statistic are employed for both training objectives and diagnostics, with the correspondence between kernel width and smoothing guiding model selection procedures (Rustamov, 2019).
References:
(Ebner et al., 2021) "Bahadur efficiencies of the Epps--Pulley test for normality" (Ebner et al., 2021) "On the eigenvalues associated with the limit null distribution of the Epps-Pulley test of normality" (Rustamov, 2019) "Closed-form Expressions for Maximum Mean Discrepancy with Applications to Wasserstein Auto-Encoders"