W-SCOMP: Weighted Sequential OMP
- W-SCOMP is a greedy algorithm for noiseless group testing that incorporates weighted scoring to enhance the identification of defective items.
- It iteratively selects candidates by updating scores based on unresolved positive tests, penalizing ambiguity and improving signal clarity.
- Empirical evaluations show that W-SCOMP offers modest gains in SNR and reduces the number of tests needed compared to traditional SCOMP.
Weighted Sequential Combinatorial Orthogonal Matching Pursuit (W-SCOMP) is a greedy decoding algorithm developed for noiseless non-adaptive group testing. Group testing aims to efficiently identify a small set of defective items from a larger population using pooled test designs. W-SCOMP optimizes the detection of defectives by incorporating a weighted scoring scheme into the classical Sequential Combinatorial Orthogonal Matching Pursuit (SCOMP) decoder, yielding improved signal-to-noise ratio (SNR) and modest but consistent empirical gains, particularly in regimes where recovery is information-limited. The development, structure, and performance guarantees of W-SCOMP are anchored in both combinatorial group testing frameworks and the broader class of weighted greedy methods such as those synthesized from Simultaneous Orthogonal Matching Pursuit with Noise Stabilization (SOMP-NS) (Determe et al., 2015, Franco-Vivo, 12 Jan 2026).
1. Foundation and Problem Setting
The W-SCOMP decoder operates within the classical combinatorial group testing model characterized by the following:
- Population and Defectives: Given items, a small unknown set (with ) is defective. The indicator vector is , where iff .
- Test Matrix: Tests are defined by , with if item is included in test .
- Non-Adaptive Noiseless Model: All tests are run in parallel. The outcome vector is given by
so if any defective is present in test .
- Priors and Designs: Typically, the defective set is drawn uniformly at random. Standard designs include the Bernoulli design (each ) and constant/near-constant tests-per-item schemes (Franco-Vivo, 12 Jan 2026).
2. Weighted Scoring: The Core Mechanism
The fundamental innovation in W-SCOMP is the replacement of the unit-based scoring function in SCOMP with a weighted contribution reflecting test ambiguity. After initial elimination using COMP or Definite Defectives (DD), the following concepts arise:
- Definite Nondefectives (DND): Items appearing in any negative test.
- Potential Defectives (PD): Set minus items identified as definite defectives.
- Unresolved Positive Tests (): Positive tests not yet explained by current defective estimates.
- Row-Weight:
for each unresolved positive test . This quantifies the number of remaining plausible defectives in a pool.
- Per-Test Contribution: For candidate in test ,
where typically . This yields each item's aggregate score:
This weighting penalizes items present in ambiguous (high ) tests and prioritizes those that better resolve unexplained positives.
3. Algorithmic Structure and Pseudocode
W-SCOMP proceeds as a sequential greedy decoder, iteratively explaining positive tests through candidate selection:
- Initialization:
- DND and PD sets identified via negative tests.
- Definite defectives collected from singleton pools if any.
- Initial estimate .
- Iterative Selection:
- While (unexplained positives) and are nonempty:
- For each , compute .
- For each , calculate .
- Add to .
- Remove tests in where .
- Update .
- While (unexplained positives) and are nonempty:
- Return as the estimated defective set.
The only modification from standard SCOMP is the use of the weighted score versus the count of unresolved positives (Franco-Vivo, 12 Jan 2026).
4. Theoretical Guarantees and SNR Analysis
Performance analysis leverages the framework of additive scoring rules to separate defectives from nondefectives:
- Per-Test SNR:
with , computed for (defective, nondefective). Aggregate SNR scales as .
- Error Bounds:
- Chebyshev: .
- Gaussian approximation: .
- Per-Test Design: In Bernoulli design (with ), closed-form are provided in Propositions 2–5 (Franco-Vivo, 12 Jan 2026). The key result is:
so W-SCOMP is never dominated by SCOMP on SNR grounds.
- Sample Complexity: To guarantee vanishing error rate (),
and with under Bernoulli design,
5. Computational Complexity
W-SCOMP retains the computational simplicity of greedy decoders:
- Initial DND and DD: .
- Per Iteration: Computing and each takes ; with up to selections, the practical runtime is , which can be reduced to with sparse data structures (Franco-Vivo, 12 Jan 2026).
6. Comparative Performance and Empirical Validation
Theoretical and empirical evaluations highlight the following:
| Algorithm | Tests for |
|---|---|
| COMP | |
| DD | |
| SCOMP | |
| W-SCOMP |
- W-SCOMP consistently achieves the same probability of exact recovery as SCOMP but with approximately 5 fewer tests (for , , Bernoulli design).
- Both false-positive and false-negative rates decline more rapidly for W-SCOMP (Franco-Vivo, 12 Jan 2026).
- Set-similarity metrics (Jaccard, -score) exhibit a small but robust edge for W-SCOMP.
- The average misclassification difference is nonnegative, confirming uniform noninferiority.
- All empirical findings are robust to variations in test design and number of defectives.
7. Broader Context and Extensions
W-SCOMP exemplifies a class of algorithms leveraging weighted joint selection rules to optimize combinatorial recovery under uncertainty. The foundational analysis for such weighted sequential/combinatorial OMP variants, including noise-stabilized frameworks like SOMP-NS (Determe et al., 2015), demonstrates a universal template:
- Introduce nonnegative per-measurement or per-test weights in the selection score.
- Propagate these weights into subsequent updates.
- Redo theoretical analysis (e.g., via restricted isometry property or exact recovery criterion) using corresponding weighted quantities.
- Choose weights according to statistical or information-theoretic optimality criteria (e.g., inverse-variance for MMV, ambiguity-based for group testing).
A plausible implication is that similar weighted greedy principles can be fruitfully extended to new combinatorial recovery settings by adapting the weight selection to the specifics of the underlying statistical model and measurement process (Determe et al., 2015, Franco-Vivo, 12 Jan 2026).