Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

On the Second-Order Asymptotics of the Hoeffding Test and Other Divergence Tests (2403.03537v2)

Published 6 Mar 2024 in cs.IT and math.IT

Abstract: Consider a binary statistical hypothesis testing problem, where $n$ independent and identically distributed random variables $Zn$ are either distributed according to the null hypothesis $P$ or the alternative hypothesis $Q$, and only $P$ is known. A well-known test that is suitable for this case is the so-called Hoeffding test, which accepts $P$ if the Kullback-Leibler (KL) divergence between the empirical distribution of $Zn$ and $P$ is below some threshold. This work characterizes the first and second-order terms of the type-II error probability for a fixed type-I error probability for the Hoeffding test as well as for divergence tests, where the KL divergence is replaced by a general divergence. It is demonstrated that, irrespective of the divergence, divergence tests achieve the first-order term of the Neyman-Pearson test, which is the optimal test when both $P$ and $Q$ are known. In contrast, the second-order term of divergence tests is strictly worse than that of the Neyman-Pearson test. It is further demonstrated that divergence tests with an invariant divergence achieve the same second-order term as the Hoeffding test, but divergence tests with a non-invariant divergence may outperform the Hoeffding test for some alternative hypotheses $Q$. Potentially, this behavior could be exploited by a composite hypothesis test with partial knowledge of the alternative hypothesis $Q$ by tailoring the divergence of the divergence test to the set of possible alternative hypotheses.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (18)
  1. W. Hoeffding, “Asymptotically optimal tests for multinomial distributions,” The Annals of Mathematical Statistics, pp. 369–401, Apr. 1965.
  2. V. Y. Tan, “Asymptotic estimates in information theory with non-vanishing error probabilities,” Foundations and Trends® in Communications and Information Theory, vol. 11, no. 1-2, pp. 1–184, 2014.
  3. M. Feder and N. Merhav, “Universal composite hypothesis testing: A competitive minimax approach,” IEEE Transactions on Information Theory, vol. 48, no. 6, pp. 1504–1517, Jun. 2002.
  4. O. Zeitouni, J. Ziv, and N. Merhav, “When is the generalized likelihood ratio test optimal?” IEEE Transactions on Information Theory, vol. 38, no. 5, pp. 1597–1602, Sep. 1992.
  5. J. Unnikrishnan, D. Huang, S. P. Meyn, A. Surana, and V. V. Veeravalli, “Universal and composite hypothesis testing via mismatched divergence,” IEEE Transactions on Information Theory, vol. 57, no. 3, pp. 1587–1603, Mar. 2011.
  6. P. Boroumand and A. Guillén i Fàbregas, “Mismatched binary hypothesis testing: Error exponent sensitivity,” IEEE Transactions on Information Theory, vol. 68, no. 10, pp. 6738–6761, Oct. 2022.
  7. ——, “Composite Neyman-Pearson hypothesis testing with a known hypothesis,” in 2022 IEEE Information Theory Workshop (ITW), Mumbai, India, Nov. 2022, pp. 131–136.
  8. S. Watanabe, “Second-order optimal test in composite hypothesis testing,” in Proc. 2018 International Symposium on Information Theory and Its Applications (ISITA), Singapore, Oct. 2018, pp. 722–726.
  9. M. Gutman, “Asymptotically optimal classification for multiple tests with empirically observed statistics,” IEEE Transactions on Information Theory, vol. 35, no. 2, pp. 401–408, Mar. 1989.
  10. Y. Li and V. Y. F. Tan, “Second-order asymptotics of sequential hypothesis testing,” IEEE Transactions on Information Theory, vol. 66, no. 11, pp. 7222–7230, Nov. 2020.
  11. S. S. Wilks, “The large-sample distribution of the likelihood ratio for testing composite hypotheses,” The Annals of Mathematical Statistics, vol. 9, no. 1, pp. 60–62, Mar. 1938.
  12. S. Watanabe, “Neyman-Pearson test and Hoeffding test,” Jan. 2018, slides, Bombay Information Theory Seminar (BITS). [Online]. Available: https://drive.google.com/file/d/1kj_lU6hkaHx8iKLOJ_9s81LrI0NglIVy/view
  13. L. L. Campbell, “An extended Čencov characterization of the information metric,” Proceedings of the American Mathematical Society, vol. 98, no. 1, pp. 135–141, Sep. 1986.
  14. S.-I. Amari and A. Cichocki, “Information geometry of divergence functions,” Bulletin of the Polish Academy of Sciences. Technical Sciences, vol. 58, no. 1, pp. 183–195, 2010.
  15. T. R. Read, “Closer asymptotic approximations for the distributions of the power divergence goodness-of-fit statistics,” Annals of the Institute of Statistical Mathematics, vol. 36, pp. 59–69, Dec. 1984.
  16. J. K. Yarnold, “Asymptotic approximations for the probability that a sum of lattice random vectors lies in a convex set,” The Annals of Mathematical Statistics, pp. 1566–1580, Oct. 1972.
  17. V. V. Ulyanov and V. N. Zubov, “Refinement on the convergence of one family of goodness-of-fit statistics to chi-squared distribution,” Hiroshima Mathematical Journal, vol. 39, no. 1, pp. 133–161, Mar. 2009.
  18. A. Alexanderian, “Some notes on asymptotic theory in probability,” Apr. 2015, technical note. [Online]. Available: https://aalexan3.math.ncsu.edu/articles/asymp-final.pdf

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com