Papers
Topics
Authors
Recent
2000 character limit reached

Post-Double-Selection LASSO

Updated 15 December 2025
  • The paper introduces a double-selection procedure that simultaneously selects covariates for both the outcome and regressor to mitigate omitted variable bias.
  • PDS-LASSO utilizes Neyman-orthogonal moments to achieve √n-consistent and asymptotically normal estimates in high-dimensional settings.
  • Empirical evidence and simulations reveal its effectiveness in stochastic-frontier analysis and many-instrument scenarios compared to single-selection methods.

Post-Double-Selection LASSO (PDS-LASSO) is a variable-selection and estimation framework designed for high-dimensional linear models, particularly when inference about a low-dimensional parameter is required in the presence of potentially many nuisance covariates. The method augments standard LASSO selection by conducting variable selection for both the outcome and the regressor of interest, followed by unpenalized estimation on the union of selected controls. This double selection, paired with Neyman-orthogonal moments, yields √n-consistent, asymptotically normal inference and mitigates omitted variable bias in settings where the number of covariates may exceed the number of observations (Parmeter et al., 20 May 2025, Hué et al., 26 Nov 2025, Chiang et al., 2019).

1. Model Framework and Identifying Conditions

PDS-LASSO is posed within a high-dimensional linear regression or partially linear model where inference is required for a low-dimensional structural parameter (e.g., a treatment effect or a coefficient on a key input) in the presence of high-dimensional controls:

  • Standard partially linear model:

yi=diθ+xiβy+εi, di=xiγ+ηi,y_i = d_i\,\theta + x_i^\top\beta^y + \varepsilon_i,\ d_i = x_i^\top\gamma + \eta_i,

where yiy_i is the scalar outcome, did_i is the regressor of interest, xiRpx_i \in \mathbb{R}^p is the high-dimensional control vector, and θ\theta is the low-dimensional parameter of interest (Hué et al., 26 Nov 2025). In stochastic-frontier settings, the model generalizes to

yi=β0+xiβ+ziδ+εi,εi=viui,y_i = \beta_0 + x_i^\top\beta + z_i^\top\delta + \varepsilon_i, \quad \varepsilon_i = v_i - u_i,

where xix_i are primary inputs and ziz_i are high-dimensional environmental (inefficiency-shifting) covariates (Parmeter et al., 20 May 2025).

  • Assumptions:
  1. Approximate sparsity: Only a small subset of regressors are relevant: βy0+γ0n\|\beta^y\|_0 + \|\gamma\|_0 \ll n, or more generally, δ0\delta_0 is ss-sparse in zz variables (sns \ll n).
  2. Restricted Eigenvalue (RE)/Compatibility: The Gram matrix of controls must satisfy compatibility to ensure identifiability and control the LASSO error (Hué et al., 26 Nov 2025).
  3. Exogeneity: Errors have mean zero conditional on controls; for stochastic-frontier, E[vixi,zi]=0E[v_i|x_i,z_i]=0 and E[uixi,zi]=E[ui]E[u_i|x_i,z_i]=E[u_i] (Parmeter et al., 20 May 2025).
  4. Moments and Tails: Variables and errors are sub-Gaussian/bounded moments of sufficiently high order.
  5. Orthogonality/Neyman-Orthogonality: The score function used in final estimation must be first-order insensitive to small estimation error in nuisance parameters (Parmeter et al., 20 May 2025).

2. The PDS-LASSO Algorithm

PDS-LASSO proceeds as follows:

  1. First Stage (LASSO for dd):

γ^=argminγ12ni=1n(dixiγ)2+λγ1\hat{\gamma} = \arg\min_{\gamma} \frac{1}{2n}\sum_{i=1}^n (d_i - x_i^\top\gamma)^2 + \lambda \|\gamma\|_1

Select S^d={j:γ^j0}\hat S_d = \{j : \hat{\gamma}_j \neq 0\} (Hué et al., 26 Nov 2025).

  1. Second Stage (LASSO for yy):

α^=argminα12ni=1n(yixiα)2+λα1\hat{\alpha} = \arg\min_{\alpha} \frac{1}{2n}\sum_{i=1}^n (y_i - x_i^\top\alpha)^2 + \lambda \|\alpha\|_1

Select S^y={j:α^j0}\hat S_y = \{j : \hat{\alpha}_j \neq 0\}.

  1. Union of Selected Controls:

S^=S^dS^y\hat S = \hat S_d \cup \hat S_y

This ensures that any variable affecting either the regressor of interest or the outcome is included.

  1. Final Estimation (Post-LASSO OLS/COLS):

(θ^,β^S^)=argminθ,βS^12ni=1n(yidiθxi,S^βS^)2(\hat{\theta}, \hat{\beta}_{\hat S}) = \arg\min_{\theta, \beta_{\hat S}} \frac{1}{2n}\sum_{i=1}^n \left(y_i - d_i\theta - x_{i, \hat S}^\top \beta_{\hat S}\right)^2

For stochastic-frontier models, analogous three-block selection and estimation steps are performed, including cross-fitting or sample splitting to ensure independence of selection and estimation phases (Parmeter et al., 20 May 2025).

A table summarizing the three main stages:

Stage Operation Selected Set
1. LASSO for dd dxd \sim x S^d\hat{S}_d
2. LASSO for yy yxy \sim x S^y\hat{S}_y
3. Final OLS yd+xS^y \sim d + x_{\hat{S}} S^dS^y\hat{S}_d \cup \hat{S}_y

3. Neyman-Orthogonality and Bias Elimination

The principal advantage of PDS-LASSO lies in the orthogonal construction of the final estimating equation:

  • Orthogonal Score: The resulting moment condition is first-order robust to small errors in the estimation of nuisance parameters. For the partially linear model, Neyman-orthogonality (partialling out xx in both yy and dd) ensures that specification mistakes in variable selection do not introduce leading-order bias in inference about θ\theta (Hué et al., 26 Nov 2025, Parmeter et al., 20 May 2025).
  • In stochastic-frontier models: Construction of a Neyman-orthogonal moment via partialling out zz from yy and xx precedes estimation of the frontier parameters. This approach restores √n-consistency even when dnd \gg n (Parmeter et al., 20 May 2025).
  • Mathematical statement: Neyman-orthogonality holds if

ηE[ψ(Wi;θ0,η)]η=η0=0\left. \frac{\partial}{\partial \eta} E[\psi(W_i;\theta_0,\eta)] \right|_{\eta=\eta_0} = 0

Such moment conditions yield final estimators whose asymptotic law does not depend on first-stage selection mistakes of order oP(1)o_P(1).

  • Implication: Single-selection LASSO (i.e., stepwise inclusion only for yy or dd) can leave bias of order OP(n1/2+δ)O_P(n^{-1/2+\delta}) in the coefficient of interest, whereas PDS-LASSO eliminates this to oP(1)o_P(1) (Parmeter et al., 20 May 2025).

4. Theoretical Guarantees and Inference

Under approximate sparsity (sy,sdn/logps_y,s_d \ll n/\log p), Restricted Eigenvalue assumptions, and appropriate regularization (λ(logp)/n\lambda \asymp \sqrt{(\log p)/n}):

  • √n-Consistency and Asymptotic Normality:

n(θ^θ)N(0,σθ2),\sqrt{n}(\hat{\theta} - \theta) \rightsquigarrow \mathcal{N}(0, \sigma_\theta^2),

with σθ2\sigma_\theta^2 consistently estimated via plug-in residuals from the final OLS (Hué et al., 26 Nov 2025, Chiang et al., 2019).

  • Oracle Inequalities: If the true support is contained in S^\hat S, the final estimator matches the performance (rate and limiting distribution) of the "oracle" estimator that knows the true support.
  • Uniformly Valid Inference: As the final regression is Neyman-orthogonal, post-selection inference is valid without data-snooping corrections beyond robust/clustered standard errors as dictated by the error structure (Hué et al., 26 Nov 2025, Chiang et al., 2019).
  • Extensions to Clustering: The method adapts to cluster settings by using cluster-robust standard errors, and by improving variance estimation to account for clustering in the design (Chiang et al., 2019).

5. Algorithmic Implementation and Practical Considerations

  • Penalty selection: The LASSO penalty parameter λ\lambda plays a critical role. If set too high, important covariates may be omitted (inducing finite sample bias); too low, and the estimator becomes high variance due to over-selection (Hué et al., 26 Nov 2025).
  • Sample splitting/cross-fitting: Used to ensure the independence between variable selection and final estimation, improving validity of inference (Parmeter et al., 20 May 2025).
  • Software: PDS-LASSO can be implemented via standard LASSO routines, often with post-processing for variable selection and final OLS.

6. Applications, Empirical Performance, and Comparisons

  • Stochastic-Frontier (Efficiency) Analysis: For estimating efficiency frontiers in settings with "big (wide) data," PDS-LASSO enables reliable measurement of inefficiency parameters by robustly selecting environmental covariates and avoiding spurious "no-inefficiency" artifacts seen with naive or single-selection approaches (Parmeter et al., 20 May 2025).
  • Empirical Evidence: Monte Carlo simulations and empirical studies highlight PDS-LASSO’s small bias, accurate coverage, and robustness to high-dimensional control vectors, contrast to single-LASSO and naive methods (Parmeter et al., 20 May 2025, Chiang et al., 2019).
  • Limitations: In finite samples, particularly with strongly negatively correlated controls or weak signal variables just below the LASSO threshold, the method may suffer omitted variable bias if relevant covariates are excluded in both selection steps (Hué et al., 26 Nov 2025). Other selection methods (e.g., Post-Double-Autometrics) have been proposed to address these circumstances.
  • Comparison to 2SLS: In many-instrument settings, PDS-LASSO avoids the weak-instrument and over-ID bias typical of traditional two-stage least squares, while maintaining inferential validity under high dimensionality (Parmeter et al., 20 May 2025).

7. Methodological Developments and Alternatives

  • Post-Double-Autometrics: As an inference-based alternative, Post-Double-Autometrics utilizes classical t-tests in general-to-specific algorithms to ensure higher "potency" (fewer missed confounders), with superior finite-sample performance and smaller RMSE demonstrated in empirical and Monte Carlo comparisons (Hué et al., 26 Nov 2025).
  • Post-Double Selection under Multi-way Clustering: The framework generalizes to arbitrary multi-way clustered sampling, provided cluster-robust variance estimation is employed (Chiang et al., 2019).

PDS-LASSO constitutes a core methodology for valid inference in high-dimensional linear and efficiency models under approximate sparsity, with theoretical properties and practical performance established across diverse econometric and statistical settings (Parmeter et al., 20 May 2025, Hué et al., 26 Nov 2025, Chiang et al., 2019).

Whiteboard

Follow Topic

Get notified by email when new papers are published related to Post-Double-Selection LASSO (PDS-LASSO).