Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 91 tok/s
Gemini 2.5 Pro 38 tok/s Pro
GPT-5 Medium 19 tok/s
GPT-5 High 23 tok/s Pro
GPT-4o 87 tok/s
GPT OSS 120B 464 tok/s Pro
Kimi K2 171 tok/s Pro
2000 character limit reached

Online Testing Problem Overview

Updated 7 September 2025
  • OTP is a framework for sequential hypothesis testing that dynamically adjusts significance levels to control errors in real-time data streams.
  • Recent methods like Alpha-investing, LORD, and e-value-based approaches enable adaptive control and efficient resource allocation under operational uncertainty.
  • Empirical and theoretical advances in OTP drive improvements in genomics, clinical trials, and autonomous systems by balancing statistical power with regulatory constraints.

The Online Testing Problem (OTP) encompasses a diverse set of research questions arising in scenarios where tests, hypotheses, or experimental actions must be executed and evaluated sequentially as data arrives over time. Unlike traditional batch testing, OTP requires policies and algorithms that guarantee error control, optimize statistical power, or maximize utility with resource constraints, all under operational uncertainty or evolving information. Current scholarship spans multiple domains: statistical multiple testing, sequential hypothesis testing, adaptive experimental design, resource allocation, systems testing of deep neural networks, and online scheduling. The following sections survey major facets and technical developments in OTP, drawing from recent literature.

1. Formal Definitions and Sequential Decision Setting

OTP fundamentally involves sequential decision-making under uncertainty, with observations, hypotheses, or test subjects arriving in a stream. In the statistical context, each time point tt is associated with a hypothesis HtH_t and a test statistic or pp-value PtP_t (or, in modern approaches, an ee-value EtE_t). Decisions (reject, accept, continue, or allocate resources) must be made immediately, possibly under cost or budget constraints.

Key formulations include:

  • Online multiple hypothesis testing: At each tt, controlling quantities such as the false discovery rate (FDR), marginal FDR (mFDR), or familywise error rate (FWER) by dynamically updating significance levels αt\alpha_t based on past decisions, with no access to future data or hypotheses (Robertson et al., 2022, Xu et al., 2023).
  • Sequential testing with resource allocation: Each test consumes non-renewable or replenishable budget; the goal may be to maximize discoveries subject to cumulative budget or error constraints using knapsack or dynamic programming formulations (Ao et al., 18 Feb 2024, Cao et al., 2020, Chen et al., 3 Sep 2025).
  • Testing policies with missing data: When only partial information is observed after early stopping, learning the joint distribution becomes more challenging; regret bounds scale worse than in classical learning settings, e.g., Ω(T2/3)\Omega(T^{2/3}) versus Θ(T)\Theta(\sqrt{T}) per (Chen et al., 3 Sep 2025).

2. Methodological Advances: Algorithms and Frameworks

Recent work has produced sophisticated algorithmic frameworks for OTP across several lines:

  • Alpha-investing and GAI procedures: These dynamically allocate and replenish "alpha-wealth" based on prior discoveries, balancing statistical power and error control (Robertson et al., 2022).
  • LORD, SAFFRON, ADDIS algorithms: These methods adaptively compute test levels using history-dependent sequences, candidate selection, and discarding mechanisms, especially to handle non-uniform or conservative nulls (Zrnic et al., 2018, Döhler et al., 2021).
  • E-value-based approaches: E-LOND and ULOND operate directly on e-values, ensuring FDR control under arbitrary dependence and enabling more powerful discoveries (Xu et al., 2023, Fischer et al., 22 Jul 2024).
  • Asynchronous and dependent settings: Conflict set frameworks guarantee mFDR in decentralized environments (e.g., overlapping tests, batch updates) via "shielding" strategies (Zrnic et al., 2018).
  • Online scheduling and resource allocation: In multiprocessor scheduling with on-the-fly testing decisions, randomized algorithms outperform deterministic lower bounds (Gong et al., 2023).

3. Theoretical Guarantees: Error Rates, Regret, and Optimality

OTP research focuses on simultaneous control of statistical summary measures and optimization of cumulative utility:

  • FDR/mFDR/FDX control: Procedures maintain expected FDP or margin guarantees over time, often using wealth update formulas and non-increasing threshold sequences (Robertson et al., 2022).
  • Anytime-valid inference and closed testing: Online closed testing via e-values establishes lower bounds on true discoveries uniformly over all data-adaptive rejection sets, harnessing Ville’s inequality and the closure principle (Fischer et al., 22 Jul 2024).
  • Regret minimization: In resource allocation and learning policies, regret bounds for optimal online strategies are characterized, with Ω(T)\Omega(\sqrt{T}) regret achievable in generic settings and improved logarithmic rates in buffered discrete cases (Ao et al., 18 Feb 2024). In missing data scenarios, regret lower bounds degrade to Ω(T2/3)\Omega(T^{2/3}) due to partial feedback (Chen et al., 3 Sep 2025).
  • Competitive analysis: Unified competitive frameworks determine the smallest possible ratio for online knapsack and trading, characterized by solutions to threshold-based differential inequalities (Cao et al., 2020).

4. Domain Applications: Online Testing in Practice

OTP methodologies are deployed across high-impact experimental and real-world settings:

  • Large-scale genomics and phenotyping: IMPC datasets, with tens of thousands of sequential tests, highlight the utility of adaptive and super-uniform rewarding methods for discovery control (Döhler et al., 2021, Robertson et al., 2022).
  • Platform clinical trials: Rolling introduction of treatments over time is accommodated by online FDR algorithms, supporting adaptive regulatory inference (Robertson et al., 2022).
  • Autonomous systems and DNN evaluation: Offline testing (static dataset metrics) is more optimistic than online closed-loop simulation, which reveals cumulative safety violations and requires real-time interaction with environments (Haq et al., 2019, Haq et al., 2021).
  • Online A/B/n experimentation: Adaptive allocation, best-arm identification, and dynamic hypothesis testing inform web service optimization and ad placement (Xu et al., 2022).
  • Resource-constrained anomaly detection: Bayesian knapsack policies maximize anomaly discoveries in time-series monitoring, such as NYC taxi passenger flows, with empirically validated error control (Ao et al., 18 Feb 2024).

5. Challenges, Limitations, and Extensions

Notable limitations and open challenges in OTP research include:

  • Dependence structures: Controlling error under unknown or complex dependence remains nontrivial, necessitating conservative or adaptive correction factors (Zrnic et al., 2018, Xu et al., 2023, Fischer et al., 22 Jul 2024).
  • Power/conservatism trade-offs: Increased shielding or conservative error control (e.g., large conflict sets) inevitably decrease discovery power, and there is a fundamental trade-off between adaptivity (speed) and power (Zrnic et al., 2018, Robertson et al., 2022).
  • Missing data: For policies with missing outcome-dependent reward, minimax regret lower bounds are elevated, making sample-efficient learning harder (Chen et al., 3 Sep 2025).
  • Rule-based hybridization: Attempts to exploit offline test outcomes to prune online testing scenarios have not yielded robust predictive rules in autonomous systems (Haq et al., 2021).
  • Contextual extensions: Leveraging context in adaptive sampling for more efficient exploration and inference is a prospective future direction (Xu et al., 2022).

6. Key Formulas, Quantitative Bounds, and Algorithmic Structures

OTP research employs precise mathematical frameworks and update formulas, among which:

  • FDR and FDP: FDP(T)=V(T)R(T)1\mathrm{FDP}(T) = \frac{V(T)}{R(T) \vee 1}, FDR(T)=E[FDP(T)]\mathrm{FDR}(T) = \mathbb{E}[ \mathrm{FDP}(T) ]
  • Alpha-wealth update: W(t)=W(t1)ϕt+RtψtW(t) = W(t-1) - \phi_t + R_t \psi_t
  • Online level for e-LOND: αt(eLOND)=αγt(Rt1+1)\alpha_t^{(\mathrm{e-LOND})} = \alpha \cdot \gamma_t \cdot (|\mathcal{R}_{t-1}| + 1)
  • Regret lower bound for OTP with missing data: Ω(T2/3)\Omega(T^{2/3}) and upper bound via Explore-Then-Commit: O~(T2/3)\tilde{O}(T^{2/3}) (Chen et al., 3 Sep 2025)
  • Dynamic programming for Bayesian knapsack: h(t,B)=E[max{h(t+1,B),1+h(t+1,Ba(t))}]h(t, B) = \mathbb{E}[ \max\{h(t+1,B), 1 + h(t+1, B - a^{(t)})\} ] (Ao et al., 18 Feb 2024)
  • Product-based e-value for intersection hypothesis: WIt=iI{1,,t}EiW_I^t = \prod_{i \in I \cap \{1, \ldots, t\}} E_i, intersection test: ϕI=1{tI such that WIt1/α}\phi_I = 1\{\exists\, t \in I \text{ such that } W_I^t \geq 1/\alpha\} (Fischer et al., 22 Jul 2024)

7. Significance and Cross-Disciplinary Impact

The Online Testing Problem is central to statistical inference and sequential decision-making in data-rich and dynamically evolving environments. Advances in adaptive error control, regret minimization, martingale-based inference, and resource allocation have enabled reliable scientific discovery, regulatory validation, and robust deployment of automated systems. These methodologies provide operational solutions for biomedicine, manufacturing, computer systems, and cyber-physical security—grounded by rigorous theoretical analysis and supported by empirical validation in high-throughput and safety-critical settings. The ongoing research directions suggest continuing development of flexible, scalable, and context-sensitive OTP frameworks across disciplines.