Orthogonal Matching Pursuit (OMP)
- Orthogonal Matching Pursuit (OMP) is a greedy algorithm for sparse recovery that iteratively selects the column most correlated with the current residual.
- The method employs Residual Ratio Thresholding (RRT) as a tuning-free stopping rule, eliminating the need for prior knowledge of sparsity or noise levels.
- OMP and its RRT adaptation have proven effective in compressed sensing, imaging, and outlier detection, offering robust support recovery under theoretical guarantees.
Orthogonal Matching Pursuit (OMP) is a canonical greedy algorithm designed for support recovery and sparse approximation in high-dimensional linear regression and compressed sensing. OMP iteratively identifies the active support of a sparse vector by sequentially selecting the column of the design or sensing matrix that is maximally correlated with the current residual, updating the support estimate, and recomputing the least-squares fit at each step. This mechanism underpins a wide array of theoretical results and practical successes in sparse signal processing, statistical learning, imaging, and related domains.
1. Algorithmic Framework and Standard OMP Procedure
OMP operates on the linear model
where has unit-norm columns, is assumed -sparse ( for the support ), and . OMP is initialized with residual and support , proceeding by:
- At iteration , select ;
- Update ;
- Compute the restricted least-squares solution , setting the complement to zero;
- Update the residual ;
- Iterate until a stopping criterion is satisfied.
Common stopping conditions require knowledge of either (fixed number of steps) or noise variance (residual norm below a threshold).
2. Motivations for Tuning-Free Recovery
In practical scenarios, neither the sparsity level nor the noise variance are typically available, and attempts to estimate them can be unreliable or computationally intensive, e.g., via cross-validation or stability selection. Conventional OMP thus depends on prior knowledge that is not robustly attainable, limiting its utility in many real-world signal and model selection problems (Kallummil et al., 2018).
This motivates the development of data-adaptive methodologies capable of support recovery and error control without such prior statistical information.
3. Residual Ratio Thresholding (RRT): Principle and Implementation
Residual Ratio Thresholding (RRT) introduces an alternative to classical stopping criteria. Instead of monitoring the residual norm itself,
RRT tracks the ratio
with . The iteration is halted at the maximal for which falls below a non-adaptive threshold , where
and denotes the CDF of the Beta distribution.
The RRT-OMP procedure is:
- Run standard OMP up to steps, recording ;
- For each , compute ;
- Set , or increase until non-empty;
- Report , .
Parameters are typically chosen as and .
4. Support Recovery Guarantees and Sample Complexity
Assuming the standard restricted isometry constant (RIC) condition , RRT-OMP achieves deterministic recovery analogously to OMP with known or , up to a mild additional SNR requirement in finite samples. Specifically,
- Finite sample result (Theorem 3): If
with , then .
- Asymptotic regime (Theorem 4): When , , and slowly (e.g., ), , so the SNR slack vanishes, and the procedure is large-sample consistent under the same RIC as optimal OMP.
- High-SNR error control (Theorem 6): As , the probability of missed discoveries , and the false-positive rate (and overall error) is bounded by .
5. Empirical Performance and Computational Considerations
Comprehensive numerical simulations validate the theoretical claims:
- In both moderate and small-sample regimes, RRT using or closely tracks the -error and support recovery rate of OMP with known or , while outperforming OMP+CV and low-complexity alternatives such as LAT.
- Empirical probability of exact recovery versus SNR is similar to oracle OMP, and the high-SNR false-alarm rate is always bounded by .
- On real data for outlier detection (Stack-loss, AR2000, Stars, Brain–Body–Weight), RRT identifies the same key outliers as classical robust methods with drastically lower computational burden.
Regarding implementation, running OMP up to steps requires at most arithmetic operations, only marginally exceeding the standard (oracle) variant, and obviates any need for cross-validation or expensive grid search strategies.
6. Relation to Other Greedy and Convex Procedures
RRT as a stopping rule is applicable to any greedy, monotonic procedure with nested supports (e.g., OMP, OLS, thresholded variants). Its operational simplicity (two hyperparameters , , both with recommended default choices) and tuning insensitivity make it attractive for settings where parameter-free or plug-and-play solutions are needed.
The additional SNR requirement over optimal OMP is negligible in the high-dimensional limit () and is, in practice, offset by a significant gain in reliability and reduced computational overhead compared to cross-validation or stability-selection-based selection.
Extension of such data-driven stopping rules to algorithms with non-monotonic support patterns (e.g., LASSO, Subspace Pursuit, CoSaMP) remains an open problem.
7. Implications for Practice and Open Challenges
Residual Ratio Thresholding transforms OMP into a data-driven sparse recovery framework requiring no prior model knowledge, while retaining the structural guarantees derived from RIC and achieving optimal support recovery up to an asymptotically negligible SNR surplus.
Key practical recommendations:
- Set , ; in most applications, asymptotic “tuning-free” operation is achieved over broad ranges.
- The method is directly applicable to high-dimensional regression, variable selection, signal processing, and any context demanding sparse recovery.
- The lack of reliance on parameter selection not only simplifies deployment but also improves robustness against user mis-specification and model uncertainty.
Further research is required to generalize RRT principles to non-monotone greedy algorithms and convex relaxations, as well as to extend empirical support in regimes with ultrahigh-dimensional or highly structured design matrices.