Bounded Shape-Constrained Lasso Estimator
- The paper introduces a penalized regression estimator that integrates sparsity and shape constraints with explicit range (box) and sum restrictions.
- It employs fused-lasso and nearly-isotonic penalties to promote blockwise structure and monotonicity, yielding optimal asymptotic properties under DAG-induced orders.
- The estimator extends to discrete distribution estimation and is computed efficiently using quadratic programming and ADMM strategies.
A bounded shape-constrained lasso-type estimator is a penalized regression or signal recovery procedure that enforces both sparsity and shape constraints—such as monotonicity, unimodality, or other graph-induced order restrictions—while subjecting the estimates to explicit range (box) and total-sum constraints. This methodology is particularly relevant for estimation over discrete structures (e.g., signals indexed by vertices of a directed acyclic graph, or probability mass functions) where one wishes to combine the adaptivity and regularization of lasso-type methods with global structural properties. The theoretical formulation covers general fused and nearly-isotonic penalties, with detailed asymptotic characterization, numerical strategies, and key applications in nonparametric discrete distribution estimation and constrained function inference.
1. Statistical Model and Problem Formulation
Let the underlying signal be , indexed by a finite set . A partial order on is encoded by a directed acyclic graph (DAG) , with oriented incidence matrix for . The base estimator is assumed to satisfy
for some random vector with nondegenerate law. The target is a refined correction of under shape constraints and penalties.
The constrained estimator is defined as
subject to
Here, is the fused-lasso (total variation) penalty and is the nearly-isotonic penalty ()—both promoting blockwise structure and monotonicity. Notably, the probability-type constraints (sum and range) are redundant: the unconstrained minimization with these penalties alone yields estimators that automatically respect these box and sum conditions (Pastukhov, 18 Nov 2025).
2. Penalty Structure and Tuning Parameter Asymptotics
The estimator incorporates two types of L1 penalties:
- Fused-lasso penalty: , encouraging blockwise constant fits and edge sparsity.
- Nearly-isotonic penalty: , which promotes monotonicity with respect to the partial order.
Tuning parameters are scaled as:
to ensure neither penalty dominates the stochastic fluctuations at the rate.
3. Asymptotic Distribution and Limiting Law
Under regularity and the scaling above, the asymptotic law is as follows. Define
where for noise , \begin{align*} V(w) = &-2\psiT w + |w|22 + \lambdaF_0 \sum{(i,j)\in E} (w_i - w_j) \mathrm{sign}(\beta{\circ}_i - \beta{\circ}_j) 1{\beta{\circ}_i \neq \beta{\circ}_j} \ & + \lambdaF_0 \sum_{(i,j)\in E} |w_i - w_j| 1{\beta{\circ}_i = \beta{\circ}_j} \ & + \lambda{NI}_0 \sum_{(i,j)\in E} (w_i - w_j) 1{\beta{\circ}_i > \beta{\circ}_j} + \lambda{NI}_0 \sum_{(i,j)\in E} |w_i - w_j|_{+} 1{\beta{\circ}_i = \beta{\circ}_j}. \end{align*} Equivalently, the limit is characterized as the unique minimizer of
This describes the distributional limit as a penalized projection of the base noise onto the space of allowable shapes, preserving the convergence rate of the original estimator (Pastukhov, 18 Nov 2025).
4. Specialization: Nearly-Isotonic Without Fusion Penalty
If and , and is already isotonic under , the asymptotic law decomposes by maximal blocks on which is constant. The limit is the concatenation of blockwise nearly-isotonic regressions:
where each
mirrors classical isotonic regression limit theory and corroborates known blockwise decoupling (Pastukhov, 18 Nov 2025).
5. Discrete Distribution Estimation on Directed Acyclic Graphs
For the estimation of discrete probability mass functions under order constraints, the methodology applies directly. Letting denote the empirical pmf over and representing the targeted shape (e.g., unimodality, multivariate monotonicity), the penalized estimator is
This estimator enforces nonnegativity, sum, and range constraints intrinsically, with the limit law matching that of the general signal case. Notably, if is empirical, , yielding asymptotically valid confidence bands under order restrictions (Pastukhov, 18 Nov 2025).
6. Key Proof Techniques and Theoretical Structure
Principal ingredients in the distributional analysis are:
- Redundancy of box and sum constraints is demonstrated by partitioning into fused constant regions and leveraging first-order optimality—region means stay within observed ranges and the total sum is preserved (Theorem 2.1).
- The limit transition employs epiconvergence: random processes are shown to converge in distribution with convexity and unique minimizer, enabling the application of Geyer's epiconvergence theorem for convergence in argmin distribution (Theorem 2.2).
- For and isotonic , the absence of fusion penalty in the limit across constant blocks leads to independent decoupling on each block.
- The penalized estimator retains the original estimator's rate ; the shape-constraint regularization at matched scaling does not deteriorate statistical efficiency (Pastukhov, 18 Nov 2025).
7. Algorithmic and Computational Approaches
Practical computation of bounded shape-constrained lasso-type estimators follows frameworks developed for the generalized or constrained lasso:
- Quadratic programming (for small to moderate problem sizes): recasts the L1-penalized, box, and linear constraints into a standard QP by splitting variables into positive and negative parts (Gaines et al., 2016).
- Alternating direction method of multipliers (ADMM): splits the parameter and constraint projections, efficiently handling large and arbitrary box and linear equality/inequality constraints.
- Piecewise-linear solution path algorithms: for all along a path, especially when repeated solutions are needed for model selection.
- The transformation from generalized lasso to constrained lasso further enables implementation for arbitrary penalty matrices and DAG structures (Gaines et al., 2016).
The methods automatically respect the relevant shape, sum, and range constraints under the specified penalty structures, with precise optimality system equations supporting convergence, rate, and uniqueness.
References:
- "Asymptotic Distribution of Bounded Shape Constrained Lasso-Type Estimator for Graph-Structured Signals and Discrete Distributions" (Pastukhov, 18 Nov 2025)
- "Algorithms for Fitting the Constrained Lasso" (Gaines et al., 2016)