Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
162 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Testing Self-Reducible Samplers (2312.10999v1)

Published 18 Dec 2023 in cs.DS

Abstract: Samplers are the backbone of the implementations of any randomised algorithm. Unfortunately, obtaining an efficient algorithm to test the correctness of samplers is very hard to find. Recently, in a series of works, testers like $\mathsf{Barbarik}$, $\mathsf{Teq}$, $\mathsf{Flash}$ for testing of some particular kinds of samplers, like CNF-samplers and Horn-samplers, were obtained. But their techniques have a significant limitation because one can not expect to use their methods to test for other samplers, such as perfect matching samplers or samplers for sampling linear extensions in posets. In this paper, we present a new testing algorithm that works for such samplers and can estimate the distance of a new sampler from a known sampler (say, uniform sampler). Testing the identity of distributions is the heart of testing the correctness of samplers. This paper's main technical contribution is developing a new distance estimation algorithm for distributions over high-dimensional cubes using the recently proposed sub-cube conditioning sampling model. Given subcube conditioning access to an unknown distribution $P$, and a known distribution $Q$ defined over ${0,1}n$, our algorithm $\mathsf{CubeProbeEst}$ estimates the variation distance between $P$ and $Q$ within additive error $\zeta$ using $O\left({n2}/{\zeta4}\right)$ subcube conditional samples from $P$. Following the testing-via-learning paradigm, we also get a tester which distinguishes between the cases when $P$ and $Q$ are $\varepsilon$-close or $\eta$-far in variation distance with probability at least $0.99$ using $O({n2}/{(\eta-\varepsilon)4})$ subcube conditional samples. The estimation algorithm in the sub-cube conditioning sampling model helps us to design the first tester for self-reducible samplers.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (55)
  1. An introduction to MCMC for machine learning. Machine Learning.
  2. An automated tool for rotational-xor cryptanalysis of arx-based primitives. In SITB.
  3. Testing of Horn Samplers. In AISTATS.
  4. Property testing of joint distributions using conditional samples. ACM Transactions on Computation Theory (TOCT).
  5. Complexity of High-Dimensional Identity Testing with Coordinate Conditional Sampling. In COLT.
  6. Handbook of markov chain monte carlo. Cambridge University Press.
  7. Sub-Gaussian random variables. Ukrainian Mathematical Journal.
  8. Random restrictions of high dimensional distributions and uniformity testing with subcube conditioning. In SODA.
  9. Testing bayesian networks. IEEE Transactions on Information Theory.
  10. Testing probability distributions using conditional samples. SIAM Journal on Computing.
  11. On the power of conditional samples in distribution testing. SIAM Journal on Computing.
  12. Distribution-aware sampling and weighted model counting for SAT. In AAAI.
  13. On testing of uniform samplers. In AAAI.
  14. A scalable and nearly uniform generator of SAT witnesses. In ICCAD.
  15. Constraint solving for test case generation: a technique for high-level design verification. In ICCD.
  16. Learning and testing junta distributions with sub cube conditioning. In COLT.
  17. Uniformity Testing over Hypergrids with Subcube Conditioning.
  18. Testing for concise representations. In FOCS.
  19. Concentration of Measure for the Analysis of Randomized Algorithms. Cambridge University Press.
  20. Efficient sampling of SAT solutions for testing. In ICSE.
  21. Elidan, G. 1998. Bayesian-Network-Repository. cs.huji.ac.il/w-galel/Repository/.
  22. Embed and project: Discrete sampling with universal hashing. NeurIPS.
  23. Uniform Solution Sampling Using a Constraint Solver As an Oracle. In UAI.
  24. Efficient parameter estimation of truncated boolean product distributions. In COLT.
  25. Designing samplers is easy: The boon of testers. In FMCAD.
  26. Testing Fourier Dimensionality and Sparsity. In ICALP.
  27. On computing the smallest four-coloring of planar graphs and non-self-reducible sets in P. Information Processing Letters.
  28. Huber, M. 2014. Near-linear time simulation of linear extensions of a height-2 poset with bounded interaction. Chicago Journal of Theoretical Computer Science.
  29. Huber, M. 2017. A Bernoulli mean estimate with known relative error distribution. Random Struct. Algorithms.
  30. Jerrum, M. 1998. Mathematical foundations of the Markov chain Monte Carlo method. In Probabilistic methods for algorithmic discrete mathematics.
  31. Random generation of combinatorial structures from a uniform distribution. Theoretical Computer Science.
  32. Planar graph coloring is not self-reducible, assuming P≠\neq≠ NP. Theoretical Computer Science.
  33. SharpSAT-TD Participating in Model Counting Competition 2021.
  34. Tolerant Testing of High-Dimensional Samplers with Subcube Conditioning. arXiv:2308.04264.
  35. Probabilistic preference logic networks. In ECAI.
  36. Learning Hidden Markov Models Using Conditional Samples. In COLT.
  37. Global partial orders from sequential data. In SIGKDD.
  38. On testing of samplers. NeurIPS.
  39. Constrained Sampling and Counting: Universal Hashing Meets SAT Solving. In Beyond NP, Papers from the 2016 AAAI Workshop, AAAI.
  40. Applications of SAT solvers to cryptanalysis of hash functions. In SAT.
  41. A SAT-based preimage analysis of reduced Keccak hash functions. Information Processing Letters.
  42. Convex rank tests and semigraphoids. SIAM Journal on Discrete Mathematics.
  43. Optimal partial-order plan relaxation via MaxSAT. Journal of Artificial Intelligence Research.
  44. Constraint-based random stimuli generation for hardware verification.
  45. Paninski, L. 2008. A Coincidence-Based Test for Uniformity Given Very Sparsely Sampled Discrete Data. IEEE Transactions on Information Theory.
  46. Peczarski, M. 2004. New results in minimum-comparison sorting. Algorithmica.
  47. On Scalable Testing of Samplers. NeurIPS.
  48. Testing probabilistic circuits. NeurIPS.
  49. Servedio, R. A. 2010. Testing by implicit learning: a brief survey. Property Testing.
  50. Extending SAT solvers to cryptographic problems. In SAT.
  51. A scalable scheme for counting linear extensions. In IJCAI.
  52. Counting linear extensions in practice: MCMC versus exponential Monte Carlo. In AAAI.
  53. Exact sampling of directed acyclic graphs from modular distributions. In UAI.
  54. Causal discovery via MML. In ICML.
  55. Simplifying boolean constraint solving for random simulation-vector generation. IEEE Trans. Comput. Aided Des. Integr. Circuits Syst.

Summary

We haven't generated a summary for this paper yet.