Communication-constrained hypothesis testing: Optimality, robustness, and reverse data processing inequalities
Abstract: We study hypothesis testing under communication constraints, where each sample is quantized before being revealed to a statistician. Without communication constraints, it is well known that the sample complexity of simple binary hypothesis testing is characterized by the Hellinger distance between the distributions. We show that the sample complexity of simple binary hypothesis testing under communication constraints is at most a logarithmic factor larger than in the unconstrained setting and this bound is tight. We develop a polynomial-time algorithm that achieves the aforementioned sample complexity. Our framework extends to robust hypothesis testing, where the distributions are corrupted in the total variation distance. Our proofs rely on a new reverse data processing inequality and a reverse Markov inequality, which may be of independent interest. For simple $M$-ary hypothesis testing, the sample complexity in the absence of communication constraints has a logarithmic dependence on $M$. We show that communication constraints can cause an exponential blow-up leading to $\Omega(M)$ sample complexity even for adaptive algorithms.
- Interactive inference under information constraints. IEEE Transactions on Information Theory, 68(1):502–516, 2022.
- Inference under information constraints I: Lower bounds from chi-square contraction. IEEE Transactions on Information Theory, 66(12), 2020.
- Inference under information constraints II: Communication constraints and shared randomness. IEEE Transactions on Information Theory, 66(12), 2020.
- Sketching, embedding and dimensionality reduction in information theoretic spaces. In International Conference on Artificial Intelligence and Statistics, AISTATS 2016, 2016.
- Communication lower bounds for statistical estimation problems via a distributed data processing inequality. In ACM Symposium on Theory of Computing, STOC 2016, 2016.
- Information-distilling quantizers. IEEE Transactions on Information Theory, 67(4):2472–2487, 2021.
- Binary hypothesis testing with deterministic finite-memory decision rules. In IEEE International Symposium on Information Theory, ISIT 2020, 2020.
- L. L. Cam. Asymptotic Methods in Statistical Decision Theory. Springer Series in Statistics. Springer New York, New York, NY, 1986.
- C. Canonne. Topics and techniques in distribution testing: A biased but representative sample. Foundations and Trends® in Communications and Information Theory, 19(6):1032–1198, 2022.
- The structure of optimal private tests for simple hypotheses. In ACM Symposium on Theory of Computing, STOC 2019, 2019.
- Pointwise bounds for distribution estimation under communication constraints. In Advances in Neural Information Processing Systems, NeurIPS 2021, 2021.
- T. M. Cover. Hypothesis testing with finite statistics. The Annals of Mathematical Statistics, 40(3):828–835, 1969.
- M. Charikar and A. Sahai. Dimension reduction in the ℓ1subscriptℓ1\ell_{1}roman_ℓ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT norm. In IEEE Symposium on Foundations of Computer Science, FOCS 2002, 2002.
- V. H. de la Peña and E. Giné. Decoupling. Springer New York, 1999.
- Communication and memory efficient testing of discrete distributions. In Conference on Learning Theory, COLT 2019, 2019.
- Minimax optimal procedures for locally private estimation. Journal of the American Statistical Association, 113(521):182–201, 2018.
- Optimality guarantees for distributed statistical estimation. arXiv preprint arXiv:1405.0782, 2014.
- Streaming algorithms for high-dimensional robust statistics. In International Conference on Machine Learning, ICML 2022 (To Appear), 2022.
- L. Devroye and G. Lugosi. Combinatorial Methods in Density Estimation. Springer Series in Statistics. Springer New York, New York, NY, 2001.
- C. Dwork and A. Roth. The Algorithmic Foundations of Differential Privacy. Foundations and Trends® in Theoretical Computer Science, 9(3-4):211–407, 2013.
- J. C. Duchi and R. Rogers. Lower bounds for locally private estimation via communication complexity. In Conference on Learning Theory, COLT 2019, 2019.
- Y. Dagan and O. Shamir. Detecting correlations with little memory and communication. In Conference on Learning Theory, COLT 2018, 2018.
- V. Feldman. A general characterization of the statistical query complexity. In Conference on Learning Theory, COLT 2017, 2017.
- Statistical algorithms and a lower bound for detecting planted cliques. Journal of The ACM, 64(2):8:1–8:37, 2017.
- On the complexity of random satisfiability problems with planted solutions. SIAM Journal on Computing, 47(4):1294–1338, 2018.
- G. L. Gilardoni. On the minimum f𝑓fitalic_f-divergence for given total variation. Comptes Rendus Mathematique, 343(11-12):763–766, 2006.
- Locally private hypothesis selection. In Conference on Learning Theory, COLT 2020, 2020.
- Extractor-based time-space lower bounds for learning. In ACM SIGACT Symposium on Theory of Computing, STOC 2018, 2018.
- M. Hellman and T. Cover. A review of recent results on learning with finite memory. In International Symposium on Information Theory (ISIT), pages 289–294, 1973.
- Learning with finite memory. Matematika, 17(3):137–156, 1973.
- K. J. Horadam. Hadamard Matrices and Their Applications. Princeton University Press, Princeton, N.J, 2007.
- Geometric lower bounds for distributed parameter estimation under communication constraints. IEEE Transactions on Information Theory, 67(12):8248–8263, 2021.
- P. J. Huber and V. Strassen. Minimax tests and the Neyman-Pearson lemma for capacities. The Annals of Statistics, 1(2), 1973.
- P. J. Huber. A robust version of the probability ratio test. The Annals of Mathematical Statistics, 36(6):1753–1758, 1965.
- Extremal mechanisms for local differential privacy. Journal of Machine Learning Research, 17:17:1–17:51, 2016.
- Metric structures in L1subscript𝐿1L_{1}italic_L start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT: Dimension, snowflakes, and average distortion. European Journal of Combinatorics, 26(8):1180–1190, 2005.
- J. Neyman and E. S. Pearson. On the problem of the most efficient tests of statistical hypotheses. Philosophical Transactions of the Royal Society of London. Series A, Containing Papers of a Mathematical or Physical Character, 231:289–337, 1933.
- Simple binary hypothesis testing under local differential privacy and communication constraints. In The Thirty Sixth Annual Conference on Learning Theory, pages 3229–3230. PMLR, 2023.
- Y. Polyanskiy and Y. Wu. Information theory: From coding to learning. 2023. Available at http://www.stat.yale.edu/~yw562/teaching/itbook-export.pdf.
- I. Sason. Tight bounds for symmetric divergence measures and a new inequality relating f𝑓fitalic_f-divergences. In IEEE Information Theory Workshop, ITW 2015, pages 1–5. IEEE, 2015.
- I. Sason. On f𝑓fitalic_f-divergences: Integral representations, local behavior, and inequalities. Entropy, 20(5):383, 2018.
- S. Shalev-Shwartz and S. Ben-David. Understanding Machine Learning: From Theory to Algorithms. Cambridge University Press, 2014.
- O. Shamir. Fundamental limits of online and distributed algorithms for statistical learning and estimation. In Advances in Neural Information Processing Systems, NeurIPS 2014, 2014.
- A. T. Suresh. Robust hypothesis testing and distribution estimation in Hellinger distance. In International Conference on Artificial Intelligence and Statistics, AISTATS 2021, 2021.
- Memory, communication, and statistical queries. In Conference on Learning Theory, COLT 2016, 2016.
- J. N. Tsitsiklis. Decentralized detection by a large number of sensors. Mathematics of Control, Signals, and Systems, 1(2):167–182, 1988.
- J. N. Tsitsiklis. Decentralized detection. In Advances in Statistical Signal Processing, pages 297–344. JAI Press, 1993.
- A. B. Tsybakov. Introduction to Nonparametric Estimation. Springer Series in Statistics. Springer New York, 2009.
- Minimax robust decentralized detection. IEEE Transactions on Information Theory, 40(1):35–40, 1994.
- M. J. Wainwright. High-Dimensional Statistics: A Non-Asymptotic Viewpoint. Cambridge University Press, 2019.
- A. Wald. Sequential tests of statistical hypotheses. The Annals of Mathematical Statistics, 16(2):117–186, 1945.
- Y. G. Yatracos. Rates of convergence of minimum distance estimators and Kolmogorov’s entropy. The Annals of Statistics, 13(2), 1985.
- J. Ziv and M. Zakai. On functionals satisfying a data-processing theorem. IEEE Transactions on Information Theory, 19(3):275–283, 1973.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.