Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
GPT-5.1
GPT-5.1 130 tok/s
Gemini 3.0 Pro 29 tok/s Pro
Gemini 2.5 Flash 145 tok/s Pro
Kimi K2 191 tok/s Pro
Claude Sonnet 4.5 34 tok/s Pro
2000 character limit reached

Hodrick t-statistics in Persistent Time Series

Updated 12 November 2025
  • Hodrick t-statistics are variants of the classical t-test tailored for time series models with persistent (nearly unit root) autoregression.
  • They rely on exact and numerical distributional characterizations and employ whitening to enable standard t-distribution inference.
  • CVF tests use simulation and linear programming to achieve robust size control and higher power compared to conventional methods.

Hodrick t-statistics refer to variants and generalizations of the classical Student t-statistic, specifically in the context of time series models with persistent or autocorrelated regressors, such as AR(1) processes with coefficients near unity. The central issue addressed by Hodrick t-statistics is the breakdown of the standard t-distribution’s finite-sample and asymptotic critical values in the presence of strong regressor persistence, resulting in nonpivotal distributions that depend on nuisance parameters. Recent research provides both exact distributional characterizations for the AR(1) case and algorithms for testing with approximately similar size, even when conventional simulation methods fail.

1. Failure of the Classical t-statistic under Persistent Autoregression

The classical Student t-statistic,

Tn=n(Xˉnμ0)sn,sn2=1n1i=1n(XiXˉn)2,T_n = \frac{\sqrt{n}\,(\bar X_n - \mu_0)}{s_n}, \quad s_n^2 = \frac{1}{n-1} \sum_{i=1}^n (X_i - \bar X_n)^2,

assumes that the sequence X1,,XnX_1,\dots,X_n is i.i.d. Gaussian. Under these conditions, TnT_n follows the tn1t_{n-1} distribution, because the numerator and denominator are independent (by Cochran’s theorem).

Under an AR(1) process,

Xt=μ+ϵt,ϵt=ρϵt1+σvt,vti.i.d. N(0,1),ρ<1,X_t = \mu + \epsilon_t, \quad \epsilon_t = \rho\,\epsilon_{t-1} + \sigma v_t, \quad v_t \sim \mathrm{i.i.d.}\ N(0,1),\quad |\rho|<1,

the sequence {Xt}\{X_t\} exhibits autocorrelation, never i.i.d. The covariance structure

$\Cov(X_t, X_u) = \frac{\sigma^2}{1-\rho^2}\rho^{|t-u|}$

invalidates Cochran’s factorization, so the distribution of TnT_n is a ratio of dependent variables and is not tn1t_{n-1} for any nn. This failure is especially pronounced for ρ\rho near unity, causing the asymptotic distribution of TnT_n to become nonpivotal, that is, dependent on ρ\rho (Benhamou, 2018).

2. Exact and Asymptotic Distributions of the AR(1) t-statistic

For finite samples, the AR(1) t-statistic Tn(ρ)T_n^{(\rho)} admits a formal expression as the ratio of a zero-mean normal and the Euclidean norm of a correlated normal vector: Tn(ρ)=NDT_n^{(\rho)} = \frac{N}{D} where

N=n(Xˉnμ),D=1n1i=1n(XiXˉn)2N = \sqrt{n}(\bar{X}_n - \mu), \qquad D = \sqrt{\tfrac{1}{n-1}\sum_{i=1}^n (X_i - \bar{X}_n)^2}

with X=μ1n+σϵX = \mu \mathbf{1}_n + \sigma \epsilon and ϵN(0,Ω)\epsilon \sim N(0, \Omega), Ωij=(1ρ2)1ρij\Omega_{ij} = (1-\rho^2)^{-1}\rho^{|i-j|}. NN and DD are no longer independent.

The explicit joint law of (N,D)(N, D) is multivariate normal for NN, and DD involves the norm of a projected Gaussian vector. Consequently, Tn(ρ)T_n^{(\rho)}’s marginal law is given by integrating over the joint density of (N,D)(N, D), as derived in (Benhamou, 2018). No closed-form in elementary functions exists, but distributions can be simulated or evaluated numerically.

As nn \to \infty with ρ\rho fixed, Tn(ρ)T_n^{(\rho)} converges in law to standard normal, reflecting the central limit theorem. For ρ0\rho \to 0, classical results are recovered. However, for ρ\rho near 1, dependence on ρ\rho remains significant in finite and moderate samples.

3. Modified t-statistics and Whitened Inference under AR(1)

To bypass nonpivotality for inference under AR(1), a modified statistic T~n\tilde T_n can be constructed by linear transformation: U=LX,LTL=Ω1,U = L X, \quad L^{T}L = \Omega^{-1}, yielding

T~n=n(Uˉnμ)s~n\tilde T_n = \frac{\sqrt{n}(\bar{U}_n - \mu)}{\tilde s_n}

where {Ui}\{U_i\} are i.i.d. normal, making T~n\tilde T_n exactly noncentral t-distributed, allowing for exact confidence intervals and tests (Benhamou, 2018). This whitening can also be effected via diagonalization of the circulant Toeplitz covariance using the discrete Fourier transform basis.

A plausible implication is that practitioners with sufficient knowledge of Ω\Omega (i.e., ρ\rho) and access to the data can always transform the sample so that standard t-distribution inference holds. This requires full knowledge of the covariance, which is often not feasible with highly persistent regressor processes.

4. Critical Value Function (CVF) Testing in Predictive Regressions

For inference about predictive slope coefficients in models with persistent regressors,

yt=μy+βxt1+ϵt,xt=ϕxt1+ηt,y_t = \mu_y + \beta x_{t-1} + \epsilon_t, \quad x_t = \phi x_{t-1} + \eta_t,

the limiting null distribution of the t-statistic

TT=β^β0s,T_T = \frac{\hat{\beta} - \beta_0}{s},

depends on the autoregressive parameter ϕ\phi and is nonpivotal for ϕ1\phi \approx 1 (Moreira et al., 2016).

The CVF test rejects H0:ββ0H_0: \beta \leq \beta_0 when TT(r)>KT,α(r)T_T(r) > K_{T,\alpha}(r), where

KIn,α(r)=i=1nkifβ0,ϕi(r)i=1nfβ0,ϕi(r)K_{I_n, \alpha}(r) = \frac{\sum_{i=1}^n k_i f_{\beta_0, \phi_i}(r)}{\sum_{i=1}^n f_{\beta_0, \phi_i}(r)}

with k=(k1,...,kn)k = (k_1, ..., k_n) chosen to ensure exact size α\alpha at each ϕi\phi_i in a grid InI_n. The grid can be refined to achieve uniform size control (to within ϵ\epsilon) over the nuisance parameter space.

This approach implements similarity directly using linear programming over simulated draws, so it does not rely on consistent estimation of ϕ\phi, splitting statistics, or randomization.

CVF Algorithm (summary)

Step Description Notes
1 Choose ϕ\phi grid InI_n covering region of interest e.g., [0.5,1.2][0.5, 1.2]
2 Simulate JJ draws r(j)r^{(j)} from baseline mixture f(r)f^*(r) Large JJ for precision
3 Compute score matrix Aj,i=fβ0,ϕi(r(j))/f(r(j))A_{j, i} = f_{\beta_0, \phi_i}(r^{(j)})/f^*(r^{(j)})
4 Solve linear program to find kk Imposes size α\alpha
5 Evaluate and refine as necessary Add grid points as needed

This method controls empirical size at the nominal level, and for T=100T = 100, α=10%\alpha = 10\%, achieves size within 1 percentage point for ϕ[0.5,1.5]\phi \in [0.5, 1.5] even with autocorrelated errors and trend, outperforming bootstrap and L2/Wright similar tests (Moreira et al., 2016).

5. Comparison to Conventional Methods and Power Properties

Traditional methods, including constant critical value t-tests, bootstrap, and subsampling, do not achieve uniform size over ϕ\phi in highly persistent settings. Bootstraps may under- or overestimate the critical value, and subsampling can be overly conservative or anti-conservative. The Wright L2 and Jansson & Moreira UMPCU tests use various conditioning and similarity-enforcing constructions, but can have non-monotonic power or intricate implementation (multiple integrals).

The CVF approach, by construction, yields approximately similar tests over all persistence levels considered, bridging the gap between stationary and nearly integrated regressors without user tuning of nuisance parameter estimators or block sizes. In power comparisons, CVF tests outperform L2 and UMPCU, especially in the nearly unit root regime; for example, with ϕ=1\phi=1 and c=10c=10, the CVF achieves >90%>90\% power, compared to 60%\leq 60\% for alternatives (Moreira et al., 2016).

6. Applications: Sharpe Ratio, Confidence Bounds, and Financial Time Series

In financial econometrics, the inference on the mean of an autocorrelated process directly determines the properties of Sharpe ratio estimators: S^=Xˉnsn\widehat{S} = \frac{\bar X_n}{s_n} Under AR(1) assumptions, its sampling distribution follows Tn(ρ)T_n^{(\rho)} or the whitened T~n\tilde{T}_n law, requiring exact or numerically integrated confidence intervals, rather than the classical t-distribution. Monte Carlo or transformation-based methods are necessary to reflect autocorrelation in the uncertainty quantification (Benhamou, 2018).

A plausible implication is that practitioners should avoid using off-the-shelf t-tables for Sharpe ratio inference when data exhibit serial correlation, even for large nn or moderate autocorrelation.

7. Broader Implications and Limitations

The paper of t-statistics under autocorrelation and near-unit-root persistence highlights the need for nonstandard inference procedures in time series. While whitening and CVF-based methods restore valid inferential properties, they entail computational cost and may require knowledge (or accurate estimation) of nuisance parameters (e.g., ρ\rho, ϕ\phi) or fine grid specification.

Uniformly similar tests constructed by CVF ensure robust inference across model regimes, and their linear programming-based implementation is numerically feasible for moderate nn and grids. The explicit breakdown of classical t-distribution results in persistent environments demonstrates the importance of context-aware critical value derivation in high-persistence time series analysis.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (2)
Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Hodrick t-statistics.