Papers
Topics
Authors
Recent
2000 character limit reached

Quantile Regression Forests (QRF)

Updated 1 December 2025
  • Quantile Regression Forests (QRF) are ensemble methods that nonparametrically estimate the entire conditional distribution instead of just the mean.
  • QRF employs weighted aggregations from randomized trees and proximity measures to extract conditional quantiles for uncertainty quantification and probabilistic forecasting.
  • The method is applied in diverse fields such as anomaly detection, survival analysis, and extreme value prediction while offering formal statistical guarantees and computational efficiency.

Quantile Regression Forests (QRF) are ensemble methods that extend random forests to construct nonparametric estimates of the entire conditional distribution of a response variable, not just the conditional mean. QRF provide a unified framework to estimate all conditional quantile levels with a single model for regression, probabilistic forecasting, uncertainty quantification, anomaly detection, and survival analysis, while inheriting robustness and scalability properties of classical regression forests (Li et al., 5 Aug 2024, Li et al., 2023, Papacharalampous et al., 2023, Taillardat et al., 2017, Zhou et al., 16 Oct 2024). The QRF methodology offers both empirical CDF outputs and high-resolution prediction intervals, underpinned by formal statistical theory, extensions for censored and time series data, targeted tuning for coverage, and proximity-based generalizations for computational efficiency.

1. Methodological Foundation

At the core of Quantile Regression Forests is the replacement of the point-estimate aggregation in regression forests by a weighted, nonparametric estimate of the conditional cumulative distribution function (CDF). For a training set D={(Xi,Yi)}i=1n\mathcal{D} = \{ (X_i, Y_i) \}_{i=1}^n with YiRY_i\in\mathbb{R} and XiRpX_i \in \mathbb{R}^p, and a new query point xx, the QRF computes weights {wi(x)}\{ w_i(x) \}, where

wi(x)=1Bb=1B1iLb(x)Lb(x)w_i(x) = \frac{1}{B} \sum_{b=1}^B \frac{1_{i \in L_b(x)}}{|L_b(x)|}

with Lb(x)L_b(x) denoting the set of training indices whose predictors fall in the same terminal node as xx in tree bb (Li et al., 5 Aug 2024, Li et al., 2023, Taillardat et al., 2017).

The conditional CDF estimator is

F^(yx)=i=1nwi(x)1{Yiy}\hat F(y \mid x) = \sum_{i=1}^n w_i(x) \cdot 1_{\{ Y_i \leq y \}}

and the conditional quantile at level τ(0,1)\tau\in(0,1) is obtained by inversion: Q^(τx)=inf{y:F^(yx)τ}.\hat Q(\tau \mid x) = \inf \{ y : \hat F(y \mid x) \geq \tau \}. This approach enables nonparametric estimation of the full conditional distribution with a single ensemble model.

2. Algorithmic Structure and Variants

Tree Construction and Weighting

QRF grows BB randomized regression trees on bootstrap replicates of training data. At each node, splitting proceeds by minimizing the post-split squared error, as in standard CART, or by more tailored criteria (e.g., gradient-quantile splits in gradient forests) (Taillardat et al., 2017). Each tree assigns weights inversely proportional to leaf size for observations in the same terminal node as xx (Li et al., 5 Aug 2024).

Proximity-based Quantile Estimation

A recent generalization replaces the aggregation of per-tree leaf weights with a direct computation of sample-wise proximities: prox(x,Xi)=1Bb=1B1{x,Xi in same leaf in b}\operatorname{prox}(x, X_i) = \frac{1}{B} \sum_{b=1}^B 1_{\{ x, X_i \text{ in same leaf in } b\}} These proximities serve as learned local similarity kernels that can be normalized and used analogously to standard QRF weights for CDF estimation and quantile extraction. The computational structure makes this approach advantageous for batch quantile queries (Li et al., 5 Aug 2024).

Extensions for Censoring, Multivariate, and Time Series Data

Censored QRF variants replace the outcome YiY_i by observed pairs (Yi,δi)(Y_i, \delta_i) (with δi\delta_i indicating event/censoring) and modify the estimation equation using a local Kaplan–Meier or Beran estimator for the conditional survival function (Li et al., 2020, Zhou et al., 16 Oct 2024).

In the multivariate setting, QRF-derived weights are combined with center-outward optimal transport to recover (conditional) vector quantile contours (Felix et al., 2023).

Generalized and time-series QRF (GRF, tsQRF) adapt tree-splitting to preserve "honesty" and other regularity conditions required for consistency under dependent data (Shiraishi et al., 2022).

3. Theoretical Guarantees and Statistical Properties

Quantile Regression Forests are consistent estimators of the conditional CDF under suitable regularity (trees grown deeply enough, honesty, randomization, Lipschitz regression function, etc.) (Li et al., 2023, Shiraishi et al., 2022, Nakamura et al., 28 Nov 2025). Uniform consistency is established for tsQRF under α\alpha-mixing sequences and GRF splitting (Shiraishi et al., 2022).

In high dimensions and large sample settings, bias-variance decomposition reveals a phase transition in statistical inference for variable importance via pinball loss risks, controlled by the forest subsampling rate β\beta (with snβs \asymp n^\beta) (Nakamura et al., 28 Nov 2025). For β<1/2\beta < 1/2, the QRF quantile estimator is asymptotically normal at O((n/s)1/2)O((n/s)^{-1/2}) rate; for β1/2\beta \ge 1/2, bias dominates and inference without analytic correction is invalid.

Censored QRF and GCQRF provide consistency for quantile estimation under right-censoring, using weighted quantile loss minimization adjusted for censoring via survival probabilities (Li et al., 2020, Zhou et al., 16 Oct 2024). Weak convergence is described via incomplete infinite-degree U-processes for GCQRF (Zhou et al., 16 Oct 2024).

4. Tuning, Efficiency, and Computational Complexity

Default QRF parameter settings often yield competitive performance, but targeted tuning is important for prediction interval coverage and efficiency (Berkowitz et al., 2 Jul 2025). The quantile coverage loss (QCL) objective directly optimizes the (out-of-bag) difference between empirical quantile coverage and the nominal level, substantially reducing both bias and MSE of coverage in repeated experiments compared to MSE or c-index tuning (Berkowitz et al., 2 Jul 2025).

Computationally, QRF inference cost is O(Blogn+)O(B \log n + \ell) per test point, where \ell is the number of quantile levels evaluated. Proximity-based QRF incurs O(nB+n)O(n B + n) upfront per query but enables efficient extraction of arbitrarily many quantile levels and reuse across queries (Li et al., 5 Aug 2024).

For extreme quantile levels, standard QRF cannot extrapolate beyond observed outcomes; extensions using local likelihood estimation of generalized Pareto parameters (ERF) address this limitation by fitting weighted EVT models using QRF weights (Gnecco et al., 2022, Taillardat et al., 2017).

5. Applications

QRF have been adopted in multiple domains:

  • Probabilistic forecasting of precipitation using satellite and sensor data, with rigorous assessment via quantile scoring rules and continuous ranked probability score (CRPS). QRF yields calibrated quantiles, competitive with and sometimes superior to ensemble model output statistics (EMOS) and boosting methods in spatial prediction, but lags LightGBM in speed and marginally in aggregate score (Papacharalampous et al., 2023, Taillardat et al., 2017).
  • Anomaly detection via context-specific QRF for explainable contextual anomaly scoring, leveraging locally estimated conditional quantiles and interval widths (Li et al., 2023).
  • Survival analysis with right-censored event times and high-dimensional predictors, using GCQRF and crf for robust, nonparametric estimation of quantile processes and variable importance (Zhou et al., 16 Oct 2024, Li et al., 2020).
  • Extreme value prediction, combining QRF proximity weights with local GPD likelihood for tail risk estimation (Gnecco et al., 2022).
  • Multivariate quantiles via optimal transport mapped from QRF-derived empirical measures (Felix et al., 2023).
  • Time series quantile regression in stationary and nonlinear autoregressive models, with tsQRF providing robust, volatility-sensitive estimates (Shiraishi et al., 2022).
  • Financial prediction intervals and risk estimation, including daily bond volume and asset volatility prediction, where proximity-based QRF yields tighter intervals and more efficient coverage (Li et al., 5 Aug 2024).

6. Limitations and Practical Considerations

QRFs are fundamentally restricted to the range of observed training responses and do not extrapolate for extreme quantiles unless combined with EVT or parametric tail models (Taillardat et al., 2017, Gnecco et al., 2022). Complexity–interpretability trade-offs arise: QRFs provide rich, flexible distributional estimates, but the ensemble structure is "black-box" and variable-importance inference becomes subtle in high-dimensional or highly correlated data (Nakamura et al., 28 Nov 2025, Papacharalampous et al., 2023). Variable importance estimation requires attention to bias regimes and potential analytic correction (Nakamura et al., 28 Nov 2025).

Hyperparameter tuning aligned with quantile estimation objectives (coverage, loss) is essential for predictive validity and interval width minimization (Berkowitz et al., 2 Jul 2025). For high-resolution or repeated quantile extraction, proximity-based QRF or batch strategies are preferable for computational efficiency (Li et al., 5 Aug 2024).

7. Table: Core QRF Formulas and Use-cases

Context Key Formula Papers
Conditional quantile Q^(τx)=inf{y:iwi(x)1Yiyτ}\hat Q(\tau|x) = \inf \{y: \sum_i w_i(x) 1_{Y_i \le y} \ge \tau \} (Li et al., 5 Aug 2024)
Censored QRF Q^τT(x):minq(1τ)G^(qx)iwi(x)1Yi>q\hat Q^T_\tau(x): \min_q | (1-\tau)\hat G(q|x) - \sum_i w_i(x) 1_{Y_i>q} | (Li et al., 2020)
Proximity-based QRF prox(x,Xi)\operatorname{prox}(x, X_i) as similarity weights for all quantile queries (Li et al., 5 Aug 2024)
Variable importance V^τ(S)=1ni=1n[ρτ(Yiq^τ,S)ρτ(Yiq^τ)]\hat V_\tau(S) = \frac{1}{n}\sum_{i=1}^n [\rho_\tau(Y_i-\hat q_{\tau,-S}) - \rho_\tau(Y_i - \hat q_\tau)] (Nakamura et al., 28 Nov 2025)
EVT extension (ERF) Ln(θ;x)=iwi(x)θ(Zi)L_n(\theta;x)=\sum_i w_i(x)\, \ell_\theta(Z_i) and tail quantile extrapolation (Gnecco et al., 2022)
Targeted tuning (QCL) Lτ=τ~τ, τ~=1ni1tiq^τ(xi)\mathcal{L}_\tau = |\tilde\tau - \tau|,\ \tilde{\tau}=\frac{1}{n}\sum_i 1_{t_i\le \hat{q}_\tau(x_i)} (Berkowitz et al., 2 Jul 2025)

References

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Quantile Regression Forests (QRF).