Local Asymptotic Normality (LAN) Property
- LAN is defined as the quadratic (Gaussian) approximation of the log-likelihood under local perturbations, serving as a foundation for asymptotic efficiency and minimax bounds.
- It is applied in diverse settings such as stochastic differential equations, jump processes, and high-frequency models to facilitate rigorous estimation in non-standard frameworks.
- Advanced techniques like spectral methods, orthogonalization of central sequences, and Malliavin calculus are employed to verify LAN and extract precise information in complex statistical models.
Local Asymptotic Normality (LAN) property is a fundamental conceptual and technical cornerstone in modern asymptotic statistics, characterizing when a statistical family of probability measures, under regular local perturbations of a parameter, admits a quadratic (Gaussian) approximation to the log-likelihood. This enables rigorous derivation of asymptotic minimax bounds, sharp efficiency results, and deep connections to the theory of optimal estimation in complex models such as stochastic differential equations, jump processes, high-frequency observations, and infinite-dimensional settings.
1. Formal Definition and Structural Essence
Let be a family of probability measures on a sequence of measurable spaces . The LAN property at a fixed requires the existence of a rate matrix and a positive-definite (or at least non-degenerate) Fisher information matrix , such that, for each fixed ,
where under as (Tran et al., 2013, Kohatsu-Higa et al., 2015, Brouste et al., 2016, Cai et al., 20 Oct 2025, Cai, 30 Dec 2025). The precise scaling in and the limiting depend on the stochastic model and asymptotic regime.
2. LAN in SDEs, Driven Diffusions, and Models with Jumps
LAN characterizations have been established for a wide array of continuous and discrete-time stochastic processes, including ergodic diffusions (Kohatsu-Higa et al., 2015), diffusions with jumps or jump-diffusions (Tran et al., 2013), fractional noise models (Brouste et al., 2016, Cai et al., 20 Oct 2025, Cai, 30 Dec 2025, Liu et al., 2015, Chiba, 2018), and high-frequency sampled models with infinite activity jumps (Ivanenko et al., 2014).
A canonical example is for an SDE with jumps, observed at high frequency: with the log-likelihood expansion under local alternatives : where is a properly normalized score vector arising from the increments, and (Kohatsu-Higa et al., 2015).
In the jump-diffusion framework: under regularity, one obtains a pathwise LAN expansion with the Fisher information matrix structured in terms of both continuous and jump components (Tran et al., 2013).
Jump-driven Lévy processes exhibiting local -stable behavior require nontrivial rate matrices—potentially non-diagonal if the process is asymmetric—reflecting collinearities in the likelihood expansion (Ivanenko et al., 2014).
3. High-Frequency and Semiparametric LAN: Rates, Non-Diagonality, and Information Structure
In high-frequency sampled models (e.g., fractional Brownian motion, mixed fBM, and fractional Gaussian noise), the score components for different parameters (e.g., volatility and Hurst index ) are often asymptotically collinear, necessitating non-diagonal rate matrices in the LAN expansion to achieve non-degenerate normal limits (Brouste et al., 2016, Cai, 30 Dec 2025). For mixed fBM, the central sequence must be constructed via an orthogonalization (lower-triangular) transform to ensure invertibility of the Fisher information matrix and efficiency of the LAN property (Cai, 30 Dec 2025).
For infinite-dimensional (nonparametric) diffusions, as in reflected scalar diffusions with unknown drift in , the LAN property is realized in terms of abstract Hilbert space operators. The LAN norm and information operator are defined via the score operator and its adjoint , and the limit experiment corresponds to a Gaussian shift in Hilbert space (Wang, 2018).
| Model Type | Scaling () | Key Feature |
|---|---|---|
| Ergodic diffusion | Standard rate if full obs. | |
| fBM/fGn (high-freq) | Non-diagonal , dependent on | Collinearity of scores, phase transition at |
| Jump models (Lévy) | Matrix with off-diagonal terms | Asymmetry and drift-skew interaction |
| SDE with periodicity | (shape, period) | Mixed rates, Fisher matrix block structure |
4. LAN in Fractional and Mixed Fractional Models
In models with fractional noise, including SDEs driven by (), or mixed processes , the LAN property can be established, but parameter identification and minimax bounds depend heavily on the interplay of sample frequency, the value of , and the influence of the white noise (Brouste et al., 2016, Cai, 30 Dec 2025).
In particular, for mixed fractional OU processes with , the process is a semimartingale and the likelihood can be handled via an explicit innovation decomposition using Volterra equation resolvents (Cai et al., 20 Oct 2025). The Fisher information matrix is then explicitly computable via spectral integrals involving the derivative of the log spectral density.
Degeneracy in the Fisher matrix (as or for small ) leads to a need for careful orthogonalization of central sequences, as certain directions in parameter space become unidentifiable or nearly collinear at the LAN scale (Cai, 30 Dec 2025).
5. Statistical Implications: Minimax Bounds, Efficiency, and Limit Experiments
LAN forms the theoretical foundation for sharp asymptotic lower bounds, optimality, and the structure of limiting statistical experiments. Once LAN holds, the experimental sequence is locally asymptotically equivalent to a Gaussian shift: any estimator sequence satisfies
with the maximum likelihood estimator (MLE) being asymptotically efficient (Tran et al., 2013, Kohatsu-Higa et al., 2015, Liu et al., 2015, Cai, 30 Dec 2025, Cai et al., 20 Oct 2025).
Minimax lower bounds (Hájek–Le Cam theorem) follow directly, formally characterizing the (lower) bound for any regular estimator's risk as the risk of estimation in a Gaussian shift with the same information. Contiguity of the shifted experiments (mutual absolute continuity in the LAN expansion) is pivotal for local minmax theory (Cai, 30 Dec 2025).
In specialized settings (multi-armed bandits, (Akker et al., 13 Dec 2025); mean-field interactions in McKean-Vlasov SDEs, (Maestra et al., 2022, Heidari et al., 17 Nov 2025)), the LAN property holds under specific regularity and identifiability conditions, with explicit block or composite forms for the Fisher information reflecting adaptive or interacting sampling designs.
6. Advanced Extensions: Mixed Normality, Cubic Approximations, and Monte Carlo
Beyond classical LAN, one encounters local asymptotic mixed normality (LAMN) for models with random/degenerate limiting information (especially in non-ergodic cases or under partial observation) (Fukasawa et al., 2020, Tran et al., 2013). Here, the limit experiment is a family of mixed (Gaussian with random covariance) shifts, quantified precisely by random Fisher information.
Recent advances include "Rescaled Local Asymptotic Normality" (RLAN), which expands the asymptotic quadratic approximation to hold on larger, neighborhoods, allowing for efficient cubic-maximum-likelihood estimation, and statistical efficiency even under scaling Monte Carlo error (Ning et al., 2020). The RLAN extension requires strong smoothness and moment bounds, but provides tools for valid estimation using noisy, simulation-based log-likelihood computations.
7. Methodological Tools and Proof Techniques
The establishment and verification of the LAN property relies on advanced techniques:
- Girsanov Theorem: Crucial for constructing likelihood ratios in models with drift-parameter perturbations for both standard and fractional diffusions (Tran, 2015, Liu et al., 2015, Cai et al., 20 Oct 2025).
- Malliavin Calculus: Essential for handling models with jumps or degenerate noise, yielding explicit score representations via Skorohod integrals and validating differentiability of transition densities (Kohatsu-Higa et al., 2015, Tran, 2015, Fukasawa et al., 2020, Heidari et al., 17 Nov 2025).
- Quadratic Mean Differentiability: Underpins provision of expansion rates and information matrices in general Markov and semimartingale settings (Tran et al., 2013, Maestra et al., 2022).
- Martingale Central Limit Theorem: Used throughout for proving convergence in law of score vectors in both i.i.d. and dependent/Markovian data (Cai et al., 20 Oct 2025, Kohatsu-Higa et al., 2015, Holbach, 2016).
- Spectral and Toeplitz-Circulant Methods: Deployed in models with long-memory/Toeplitz covariance (e.g., fGn, mixed fBM), to compute relevant variances and ensure the LAN norm is well-posed (Brouste et al., 2016, Cai, 30 Dec 2025).
- Ergodic Theorems: Justify convergence of empirical (information) matrices to their deterministic (Fisher) limits (Kohatsu-Higa et al., 2015, Cai et al., 20 Oct 2025).
These technical developments underpin contemporary statistical inference for stochastic dynamical systems, clarify phase transitions in estimation rates, and support efficiency in the broad field of stochastic models.