Bruss's Odds Theorem & Optimal Stopping Rule
- Bruss's Odds Theorem is a fundamental result in optimal stopping theory that defines a sum-of-odds rule to maximize the chance of stopping on the final success in a sequence.
- It uses dynamic programming and backward induction to establish a unique threshold, offering computational simplicity and clear quantitative probability bounds.
- The theorem has practical applications in secretary problems, clinical trials, and various online selection scenarios, ensuring optimal decision-making across diverse settings.
Bruss's Odds Theorem is a fundamental result in optimal stopping theory addressing sequential selection from finite sequences of independent Bernoulli trials. Its core contribution is the identification of an explicit rule—the "sum-of-odds" strategy—for maximizing the probability of successfully stopping on the very last occurrence of a specified event, which is mathematically formalized as maximizing the probability of stopping on the final "success" (i.e., the last 1) in a sequence of 0–1 (failure–success) observations. The theorem's generality and computational simplicity have established it as a central tool in both theoretical and applied sequential decision problems, including secretary problems, search theory, and clinical trials (Dendievel, 2012, Bruss, 2019).
1. Mathematical Formulation and Statement
Let be independent Bernoulli random variables on a probability space , with success probabilities , and . The objective is to select a stopping time relative to the natural filtration so as to maximize the probability of stopping exactly on the final 1: Bruss's Odds Theorem emphasizes the use of tail odds ratios: The optimal strategy is constructed as follows:
- Define the threshold .
- The optimal stopping rule is , or if no such exists.
The maximal success probability is
This threshold and win-probability formula, and their equivalence across multiple paper variants, are thoroughly established in (Kabaeva et al., 26 Nov 2025, Dendievel, 2012, Bruss, 2019, ribas, 2018, Ribas, 2018).
2. Proof Technique and Structural Properties
The optimality proof leverages dynamic programming and one-step look-ahead arguments. Central is the monotonicity of the tail odds sum , with the backward induction establishing that after a unique threshold index , immediate stopping on a success dominates continuation. At each time, the choice is between stopping immediately (with probability ) and continuing (leading to the recursive value ), showing that the policy region where it is optimal to stop comprises a single contiguous block (Dendievel, 2012, Bruss, 2019, Ribas, 2018, ribas, 2018).
3. Quantitative Success Probability Bounds
Bruss's Odds Theorem yields precise and, in many cases, sharp probabilistic bounds for the probability of success under the optimal rule:
- Upper Bound:
with equality if all odds mass is concentrated at and for .
- Lower Bounds:
- If , then , and
with equality when all equal . - If ,
with equality for for . - For , similar constructions provide bounds with sharpness achieved in limiting configurations.
The classical universal bound when follows from convexity via
and is unattainable, but approached in the i.i.d. Bernoulli limit (Kabaeva et al., 26 Nov 2025, Ribas, 2018).
4. Extensions and Generalizations
Bruss's theorem underpins an extensive literature of extensions and modifications:
- Weighted Payoffs: For weights , the threshold becomes the last index for which the weighted odds sum exceeds . The optimal stopping rule and expected reward can still be written in terms of these sums (ribas, 2018).
- Stopping on the m-th Last Success: The optimal rule generalizes to summing m-fold products of the odds, yielding analogous sum-of-multiplicative-odds thresholds (Dendievel, 2012).
- Dependent Structures: Markovian (e.g., two-state Markov chains), continuous-time (Poisson process with unknown intensity), and adversarial variants retain versions of the sum-of-odds criterion under suitably generalized conditions (Dendievel, 2012).
- Multiple Stopping Opportunities: With up to choices, recursive auxiliary odds-sums produce multiple thresholds for sequential actions (Dendievel, 2012).
5. Monotonicity, Structure, and Computational Aspects
The win-probability and threshold exhibit predictable monotonicity with respect to changes in the underlying sequence :
- If is non-increasing, the optimal value is non-increasing and threshold increments are tightly controlled.
- If is non-decreasing, thresholds and values increase.
- The unimodality of the mapping ensures the uniqueness and robustness of the threshold (Bruss, 2019). The computational recipe is recursive and efficient: backwards summing of the odds determines the threshold in time.
6. Canonical Examples and Practical Implications
The theorem provides the optimal solution for the classical secretary problem, where
and the success probability asymptotically approaches $1/e$ as . Applications include:
- Secretary and Best-Choice Problems: Derivation and justification of the celebrated $1/e$-law.
- Clinical Trials: Selection policies that minimize the risk of unnecessary further treatment after the last observed benefit (Bruss, 2019).
- Online Selection and Search: Robotic maintenance, investment timing, and other contexts where one-stop decisions must be made in real-time (Dendievel, 2012).
7. Variants, Further Bounds, and Limitations
Recent works improve lower bounds in subcritical regimes (), analyze variants such as repeating the process until a success is observed, and consider options to predict no successes. These modifications continue to guarantee nontrivial lower bounds on win probabilities, with optimal rules expressible in the sum-of-odds paradigm (Ribas, 2018).
The theorem's power is matched by clear quantitative limits: the achievable win-probability is bounded above by and below by regime-specific, often tight, analytic expressions. In all cases, sharpness is either achieved or approached asymptotically via explicit constructions (Kabaeva et al., 26 Nov 2025). The universality and robustness of the sum-of-odds method provide both theoretical insight and practical stopping heuristics that are optimal for a broad class of sequential decision problems.