Luckiness-weighted NML
- LNML is a generalized form of Normalized Maximum Likelihood that incorporates a non-negative luckiness weight, ensuring a proper minimax regret solution.
- The method regularizes models, particularly in continuous or high-dimensional settings, by modifying the likelihood with a luckiness function to avoid divergence.
- Practical applications of LNML include multivariate normal models, discrete memoryless sources, and ridge-regression-like scenarios, bridging traditional NML and Bayesian techniques.
Luckiness-weighted Normalized Maximum Likelihood (LNML) is a generalized universal distribution extending the normalized maximum likelihood (NML) to parametric models where NML is ill-defined or divergent, particularly continuous or high-capacity settings. LNML introduces a non-negative "luckiness" or weight function over the parameter space, regularizing the model and ensuring well-posedness of the minimax regret solution. LNML appears in statistical inference, coding theory, model selection, and recent advances in regularized estimation and high-dimensional settings.
1. Formal Definition and Core Properties
Given a parametric family and a sample , standard NML is defined as
If diverges (e.g., for Gaussian models), NML is not defined. LNML replaces the maximum likelihood in both numerator and denominator with a luckiness-weighted form using a weight (luckiness) function : With , LNML reduces to ordinary NML. LNML is the unique pointwise minimax solution to the regret function
so that
LNML always yields a proper (normalized) distribution assuming integrability of the weighted likelihood.
2. Minimax Regret, Asymptotics, and Interpretations
The regret of LNML under the luckiness-weighted regime is defined as
where 0. The worst-case regret is
1
so LNML achieves constant regret determined by the log normalization term. For regular (smooth) parametric families of dimension 2, the asymptotic expansion of 3 under the Laplace method is
4
where 5 is the Fisher information. This yields the same leading minimax 6 growth as NML for appropriate 7.
3. Construction and Examples in Key Parametric Families
3.1 Multivariate Normal Models
For observations 8, with 9 the standard 0-Gaussian likelihood, LNML uses a conjugate-like prior as luckiness: 1 with 2, 3, 4. This choice ensures convergence of the maximization and the normalization integrals. The resulting LNML has a closed-form: 5 with explicit formulas for 6 as weighted MAP estimators, and 7 involving special functions (multivariate Gamma).
3.2 Discrete Memoryless Sources (DMS)
With categorical probabilities 8 and luckiness 9, the LNML numerator becomes
0
for counts 1 (2), and normalization is via summing over count vectors. For 3 (“Jeffreys” luckiness), leading order regret matches NML, with a different 4 offset.
3.3 Linear Regression with 5 Luckiness
For linear regression with Gaussian errors and a ridge-like luckiness, 6, LNML in the supervised predictive version (LpNML) yields not only consistent regularization but a predictive distribution that can be computed exactly as a shifted Gaussian, blending in-sample interpolation with conservative extrapolation in under-determined cases (Bibas et al., 2022).
4. Algorithmic and Theoretical Insights
LNML density and its normalization constant typically require inner maximization and outer integration—often intractable in high dimensions. For penalized empirical risk minimization, let 7 be a penalty (luckiness), then LNML code length is
8
To address computational challenges, analytic upper bounds ("uLNML") were derived (Miyaguchi et al., 2018). Under smoothness and convexity assumptions, 9 is computable in closed-form for 0/1 penalties and is uniformly close to the true LNML code length—enabling practical parameter selection (MDL-RS) in high dimensions.
5. Role of Luckiness and Incorporation of Side Information
The choice of luckiness function 2 (or 3, 4) encodes prior beliefs, regularization, or auxiliary information:
- Priors or pseudo-priors: Conjugate-like weights encode beliefs or enforce lower bounds (e.g., for covariance matrices).
- Side information: Incidental data or null-hypothesis values can be incorporated by constructing 5 to bias estimators towards plausible regions or to allow finite regret when ordinary NML diverges.
- Regularization: 6 luckiness directly yields ridge regression behavior, regularizing hypothesis space and controlling complexity in high-capacity settings (Bibas et al., 2022).
- Statistical evidence measures: LNML enables discrimination information 7 to assess evidence for model comparison, with asymptotic calibration and robustness to multiplicity (Bickel, 2010).
6. Connections to NML, Bayesian Mixtures, and 8-NML
LNML unifies and interpolates various universal coding/prediction paradigms:
- With 9, ordinary NML is recovered.
- LNML is a limiting case of 0-NML as 1 with a prior 2 (Bondaschi et al., 2022). Mixture/Bayesian codes correspond to 3, LNML to 4.
- LNML (with appropriate luckiness) sits at the edge of the trade-off between mixture predictors and hard minimax NML, providing a uniform constant-regret bound and avoiding divergences present in unconstrained NML.
7. Practical Applications and Model Selection
LNML is particularly relevant in:
- Model selection under MDL: Explicit, finite-complexity penalty even for non-compact or continuous models, augmenting MDL-based criteria (Miyaguchi, 2017, Miyaguchi et al., 2018).
- High-dimensional penalty selection: The MDL-RS method leverages analytic uLNML to select regularization parameters efficiently, outperforming cross-validation and BIC/AIC in highly-redundant/high-dimensional regimes (Miyaguchi et al., 2018).
- Prediction under distribution shift: LNML/LpNML provides bounded, calibrated regret and improved robustness over empirical risk minimization, with improved out-of-distribution characteristics (Bibas et al., 2022).
- Robust inference and multiple comparisons: LNML-based discrimination information adapts robustly when integrating diverse side information and controls error rates in high-throughput testing (Bickel, 2010).
References:
- "Normalized Maximum Likelihood with Luckiness for Multivariate Normal Distributions" (Miyaguchi, 2017)
- "Statistical inference optimized with respect to the observed sample for single or multiple comparisons" (Bickel, 2010)
- "High-dimensional Penalty Selection via Minimum Description Length Principle" (Miyaguchi et al., 2018)
- "Alpha-NML Universal Predictors" (Bondaschi et al., 2022)
- "Beyond Ridge Regression for Distribution-Free Data" (Bibas et al., 2022)