On the Second-Order Asymptotics of the Hoeffding Test and Other Divergence Tests (2403.03537v2)
Abstract: Consider a binary statistical hypothesis testing problem, where $n$ independent and identically distributed random variables $Zn$ are either distributed according to the null hypothesis $P$ or the alternative hypothesis $Q$, and only $P$ is known. A well-known test that is suitable for this case is the so-called Hoeffding test, which accepts $P$ if the Kullback-Leibler (KL) divergence between the empirical distribution of $Zn$ and $P$ is below some threshold. This work characterizes the first and second-order terms of the type-II error probability for a fixed type-I error probability for the Hoeffding test as well as for divergence tests, where the KL divergence is replaced by a general divergence. It is demonstrated that, irrespective of the divergence, divergence tests achieve the first-order term of the Neyman-Pearson test, which is the optimal test when both $P$ and $Q$ are known. In contrast, the second-order term of divergence tests is strictly worse than that of the Neyman-Pearson test. It is further demonstrated that divergence tests with an invariant divergence achieve the same second-order term as the Hoeffding test, but divergence tests with a non-invariant divergence may outperform the Hoeffding test for some alternative hypotheses $Q$. Potentially, this behavior could be exploited by a composite hypothesis test with partial knowledge of the alternative hypothesis $Q$ by tailoring the divergence of the divergence test to the set of possible alternative hypotheses.
- W. Hoeffding, “Asymptotically optimal tests for multinomial distributions,” The Annals of Mathematical Statistics, pp. 369–401, Apr. 1965.
- V. Y. Tan, “Asymptotic estimates in information theory with non-vanishing error probabilities,” Foundations and Trends® in Communications and Information Theory, vol. 11, no. 1-2, pp. 1–184, 2014.
- M. Feder and N. Merhav, “Universal composite hypothesis testing: A competitive minimax approach,” IEEE Transactions on Information Theory, vol. 48, no. 6, pp. 1504–1517, Jun. 2002.
- O. Zeitouni, J. Ziv, and N. Merhav, “When is the generalized likelihood ratio test optimal?” IEEE Transactions on Information Theory, vol. 38, no. 5, pp. 1597–1602, Sep. 1992.
- J. Unnikrishnan, D. Huang, S. P. Meyn, A. Surana, and V. V. Veeravalli, “Universal and composite hypothesis testing via mismatched divergence,” IEEE Transactions on Information Theory, vol. 57, no. 3, pp. 1587–1603, Mar. 2011.
- P. Boroumand and A. Guillén i Fàbregas, “Mismatched binary hypothesis testing: Error exponent sensitivity,” IEEE Transactions on Information Theory, vol. 68, no. 10, pp. 6738–6761, Oct. 2022.
- ——, “Composite Neyman-Pearson hypothesis testing with a known hypothesis,” in 2022 IEEE Information Theory Workshop (ITW), Mumbai, India, Nov. 2022, pp. 131–136.
- S. Watanabe, “Second-order optimal test in composite hypothesis testing,” in Proc. 2018 International Symposium on Information Theory and Its Applications (ISITA), Singapore, Oct. 2018, pp. 722–726.
- M. Gutman, “Asymptotically optimal classification for multiple tests with empirically observed statistics,” IEEE Transactions on Information Theory, vol. 35, no. 2, pp. 401–408, Mar. 1989.
- Y. Li and V. Y. F. Tan, “Second-order asymptotics of sequential hypothesis testing,” IEEE Transactions on Information Theory, vol. 66, no. 11, pp. 7222–7230, Nov. 2020.
- S. S. Wilks, “The large-sample distribution of the likelihood ratio for testing composite hypotheses,” The Annals of Mathematical Statistics, vol. 9, no. 1, pp. 60–62, Mar. 1938.
- S. Watanabe, “Neyman-Pearson test and Hoeffding test,” Jan. 2018, slides, Bombay Information Theory Seminar (BITS). [Online]. Available: https://drive.google.com/file/d/1kj_lU6hkaHx8iKLOJ_9s81LrI0NglIVy/view
- L. L. Campbell, “An extended Čencov characterization of the information metric,” Proceedings of the American Mathematical Society, vol. 98, no. 1, pp. 135–141, Sep. 1986.
- S.-I. Amari and A. Cichocki, “Information geometry of divergence functions,” Bulletin of the Polish Academy of Sciences. Technical Sciences, vol. 58, no. 1, pp. 183–195, 2010.
- T. R. Read, “Closer asymptotic approximations for the distributions of the power divergence goodness-of-fit statistics,” Annals of the Institute of Statistical Mathematics, vol. 36, pp. 59–69, Dec. 1984.
- J. K. Yarnold, “Asymptotic approximations for the probability that a sum of lattice random vectors lies in a convex set,” The Annals of Mathematical Statistics, pp. 1566–1580, Oct. 1972.
- V. V. Ulyanov and V. N. Zubov, “Refinement on the convergence of one family of goodness-of-fit statistics to chi-squared distribution,” Hiroshima Mathematical Journal, vol. 39, no. 1, pp. 133–161, Mar. 2009.
- A. Alexanderian, “Some notes on asymptotic theory in probability,” Apr. 2015, technical note. [Online]. Available: https://aalexan3.math.ncsu.edu/articles/asymp-final.pdf