Scaled Chi Distributions (UTSCDs)
- Scaled Chi Distributions (UTSCDs) are the marginal distributions of the Euclidean norm of an n-dimensional Gaussian vector, modified via unilateral truncation at a fixed cutoff.
- Their computation utilizes closed-form expressions with incomplete gamma functions to derive PDFs, CDFs, and quantiles, enabling precise statistical calibration.
- They are crucial for applications in model fitting, goodness-of-fit testing, and extreme-value theory, especially under variance optimization constraints.
Scaled chi distributions (abbreviated UTSCDs, Unilaterally Truncated Scaled Chi Distributions, in the literature) arise as the marginal distribution of the Euclidean norm of a centered -dimensional Gaussian vector . When attention is restricted to the half-line for a fixed cutoff , the resulting truncated distributions play a central role in statistical modeling, goodness-of-fit calibration, and extreme-value theory. Recent research has determined tight bounds on their variance when a mean constraint is imposed, illuminating the structure of truncated moment spaces and offering precise methods for distributional fitting and quantile computation (Petrella, 14 Nov 2025, Das, 2024).
1. Definition and Fundamental Properties
The -dimensional radially symmetric Gaussian, , induces a marginal distribution on , known as a scaled chi distribution. Its probability density function is
Here is the degrees of freedom (dimension), the scale, and the Euler gamma function.
Unilateral truncation restricts attention to the domain for fixed , yielding the normalized truncated density
where denotes the upper incomplete gamma function.
This generalizes the central chi and half-normal distributions, with the scaled chi as a special case of a generalized chi-square quadratic form (Das, 2024). The mass beyond is .
2. Calculation of Truncated Moments
Moments of the truncated distribution quantify its central tendency and dispersion subject to the truncation: Key quantities include the truncated mean,
and the truncated second moment,
For , the truncated variance is . These formulas enable explicit computation using incomplete gamma special functions, with direct SciPy and R implementations for practical applications (Das, 2024).
3. Maximal Variance under Fixed Mean and Cutoff
For applications requiring calibration to a fixed (finite) mean on , one may optimize the variance over permissible scale and dimension . The supremum is analytically attained for (the half-normal case), with the maximal variance
where and hence .
No other choice of or achieves a larger variance under the truncated mean constraint. If is strictly positive, the maximum becomes a function of and the optimizing dimension —see next section for this dependency (Petrella, 14 Nov 2025).
4. Dependency of Variance on Dimension and Cutoff
Fixing and yields an optimization problem over . The truncated variance , viewed as a function of continuous , exhibits a unique global maximum at , which increases monotonically with cutoff (formally with ). As , , and the variance diverges, enabling arbitrarily high dispersion for vanishing cutoff when is not restricted to integer values.
Restricting to constrains the maximizer to integer values, creating a staircase in as increases. This suggests that optimal calibration for truncated variance depends sensitively on the dimensionality, with fractional outside traditional probabilistic interpretation but useful for extremal studies.
5. Computational Methods for PDF, CDF, and Quantiles
For practical evaluation, the scaled chi distribution admits closed-form expressions for PDF, CDF, and quantiles in terms of incomplete gamma functions:
- PDF: for ,
- CDF: ,
- Quantiles: , where solves .
Python, R, and Matlab example codes for these calculations were provided by Das (2024), leveraging SciPy's gammainc, gammaincinv and analogous functions, as well as an open-source Matlab toolbox (github.com/abhranildas/gx2) supporting multiple exact and approximate computational methods (series-expansion "Ruben", characteristic function inversion "Imhof", inverse FFT, and ray-tracing, with tailored precision and speed for diverse use cases) (Das, 2024).
| Software | Function names | Methods Supported |
|---|---|---|
| Python/SciPy | pdf_scaled_chi, cdf_scaled_chi | Closed-form via incomplete gamma |
| R (stats) | pdf_scaled_chi, cdf_scaled_chi | Closed-form via pchisq, dchisq |
| Matlab (Das) | gx2cdf, gx2pdf, gx2inv | ‘ruben’, ‘imhof’, ‘ifft’, ‘ray’ |
6. Calibration, Applications, and Theoretical Implications
The closed-form and optimization properties of UTSCDs position them as canonical models for statistical calibration under truncation, goodness-of-fit procedures, tail risk analysis, and simulation of radial norms. Model calibration can proceed by solving for in for given , then calculating desired moments and quantiles. The variance bounds facilitate testing and model fitting when only truncated means are observable.
A plausible implication is that the blow-up of variance for fractional in the limit provides theoretical upper bounds for dispersion under severe truncation, relevant for extremal value theory and robust statistics. The connection to generalized chi-square and quadratic forms opens avenues for further generalization, e.g., for noncentral or weighted sums.
7. Relation to Generalized Chi-Square and Quadratic Forms
Scaled chi distributions constitute a subclass of the generalized chi-square family, i.e., distributions of quadratic forms in Gaussian vectors. The generalized case, expressible as , encompasses weighted sums of (possibly noncentral) chi-square variables and normal terms. Exact computation is enabled by Imhof-type characteristic function inversion and Ruben-type series expansion; Pearson moment-matching and saddlepoint/Fourier methods provide rapid approximations with quantified error properties.
Direct specialization to , , retrieves the UTSCD structure, justifying use of incomplete gamma functions and the computational recipes previously described (Das, 2024). The relationship clarifies why scaled chi statistics are so broadly applicable, and situates the maximal variance result as a central component in the toolbox of extremal analysis and statistical calibration for truncated domains.