Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Efficient Approximation Schemes for Stochastic Probing and Prophet Problems (2007.13121v2)

Published 26 Jul 2020 in cs.DS and cs.GT

Abstract: Our main contribution is a general framework to design \emph{efficient} polynomial time approximation schemes (EPTAS) for fundamental stochastic combinatorial optimization problems. Given an error parameter $\epsilon>0$, such algorithmic schemes attain a $(1-\epsilon)$-approximation in $t(\epsilon)\cdot poly(n)$ time, where $t(\cdot)$ is some function that depends only on $\epsilon$. Technically speaking, our approach relies on presenting tailor-made reductions to a newly-introduced multi-dimensional load balancing problem. Even though the single-dimensional problem is already known to be APX-Hard, we prove that an EPTAS can be designed under certain structural assumptions, which hold for each of our applications. To demonstrate the versatility of our framework, we first study selection-stopping settings to derive an EPTAS for the Free-Order Prophets problem [Agrawal et al., EC'20] and for its cost-driven generalization, Pandora's Box with Commitment [Fu et al., ICALP'18]. These results constitute the first approximation schemes in the non-adaptive setting and improve on known {inefficient} polynomial time approximation schemes (PTAS) for their adaptive variants. Next, turning our attention to stochastic probing problems, we obtain an EPTAS for the adaptive ProbeMax problem as well as for its non-adaptive counterpart; in both cases, state-of-the-art approximability results have been inefficient PTASes [Chen et al., NIPS'16; Fu et al., ICALP'18].

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Danny Segev (18 papers)
  2. Sahil Singla (66 papers)
Citations (37)

Summary

We haven't generated a summary for this paper yet.