The Sample Complexity of Approximate Rejection Sampling with Applications to Smoothed Online Learning (2302.04658v3)
Abstract: Suppose we are given access to $n$ independent samples from distribution $\mu$ and we wish to output one of them with the goal of making the output distributed as close as possible to a target distribution $\nu$. In this work we show that the optimal total variation distance as a function of $n$ is given by $\tilde\Theta(\frac{D}{f'(n)})$ over the class of all pairs $\nu,\mu$ with a bounded $f$-divergence $D_f(\nu|\mu)\leq D$. Previously, this question was studied only for the case when the Radon-Nikodym derivative of $\nu$ with respect to $\mu$ is uniformly bounded. We then consider an application in the seemingly very different field of smoothed online learning, where we show that recent results on the minimax regret and the regret of oracle-efficient algorithms still hold even under relaxed constraints on the adversary (to have bounded $f$-divergence, as opposed to bounded Radon-Nikodym derivative). Finally, we also study efficacy of importance sampling for mean estimates uniform over a function class and compare importance sampling with rejection sampling.
- Round-optimal lattice-based threshold signatures, revisited. Cryptology ePrint Archive, Paper 2022/634, 2022. URL https://eprint.iacr.org/2022/634. https://eprint.iacr.org/2022/634.
- Local differential privacy is equivalent to contraction of ℰγsubscriptℰ𝛾\mathcal{E}_{\gamma}caligraphic_E start_POSTSUBSCRIPT italic_γ end_POSTSUBSCRIPT-divergence. arXiv preprint arXiv:2102.01258, 2021.
- Fat-shattering and the learnability of real-valued functions. In Proceedings of the seventh annual conference on Computational learning theory, pages 299–310, 1994.
- Resampled priors for variational autoencoders. In The 22nd International Conference on Artificial Intelligence and Statistics, pages 66–75. PMLR, 2019.
- Concentration inequalities, large and moderate deviations for self-normalized empirical processes. The Annals of Probability, 30(4):1576–1604, 2002.
- Efficient and near-optimal smoothed online learning for generalized linear functions. arXiv preprint arXiv:2205.13056, 2022.
- Majorizing measures, sequential complexities, and online learning. In Conference on Learning Theory, pages 587–590. PMLR, 2021.
- Smoothed online learning is as easy as statistical learning. arXiv preprint arXiv:2202.04690, 2022.
- Smoothed online learning for prediction in piecewise affine systems, 2023.
- Learnability and the vapnik-chervonenkis dimension. Journal of the ACM (JACM), 36(4):929–965, 1989.
- Prediction, learning, and games. Cambridge university press, 2006.
- The sample size required in importance sampling. The Annals of Applied Probability, 28(2):1099–1135, 2018.
- Learning bounds for importance weighting. Advances in neural information processing systems, 23, 2010.
- On rejection sampling in lyubashevsky’s signature scheme. Cryptology ePrint Archive, 2022.
- An approximate sampler for energy-based models with divergence diagnostics. Transactions on Machine Learning Research, 2022.
- Bernard D Flury. Acceptance–rejection sampling made easy. Siam Review, 32(3):474–476, 1990.
- Adaptive rejection sampling for gibbs sampling. Journal of the Royal Statistical Society: Series C (Applied Statistics), 41(2):337–348, 1992.
- Mathematical foundations of infinite-dimensional statistical models. Cambridge university press, 2021.
- Concave-convex adaptive rejection sampling. Journal of Computational and Graphical Statistics, 20(3):670–691, 2011.
- Variational rejection sampling. In International Conference on Artificial Intelligence and Statistics, pages 823–832. PMLR, 2018.
- Smoothed analysis of online and differentially private learning. Advances in Neural Information Processing Systems, 33:9203–9215, 2020.
- Oracle-efficient online learning for beyond worst-case adversaries. arXiv preprint arXiv:2202.08549, 2022a.
- Smoothed analysis with adaptive adversaries. In 2021 IEEE 62nd Annual Symposium on Foundations of Computer Science (FOCS), pages 942–953. IEEE, 2022b.
- On pairs of f𝑓fitalic_f-divergences and their joint range. IEEE Transactions on Information Theory, 57(6):3230–3235, 2011.
- The communication complexity of correlation. In Twenty-Second Annual IEEE Conference on Computational Complexity (CCC’07), pages 10–23. IEEE, 2007.
- Approximating the permanent with deep rejection sampling. Advances in Neural Information Processing Systems, 34:213–224, 2021.
- The computational power of optimization in online learning. In Proceedings of the forty-eighth annual ACM symposium on Theory of Computing, pages 128–141, 2016.
- Efficient algorithms for online decision problems. Journal of Computer and System Sciences, 71(3):291–307, 2005.
- Efficient distribution-free learning of probabilistic concepts. Journal of Computer and System Sciences, 48(3):464–497, 1994.
- Bayesian estimates of equation system parameters: an application of integration by monte carlo. Econometrica: Journal of the Econometric Society, pages 1–19, 1978.
- Nick Littlestone. Learning quickly when irrelevant attributes abound: A new linear-threshold algorithm. Machine learning, 2(4):285–318, 1988.
- Rejection sampling and noncausal sampling under moment constraints. In 2018 IEEE International Symposium on Information Theory (ISIT), pages 1565–1569. IEEE, 2018.
- ℰγsubscriptℰ𝛾\mathcal{E}_{\gamma}caligraphic_E start_POSTSUBSCRIPT italic_γ end_POSTSUBSCRIPT -resolvability. IEEE Transactions on Information Theory, 63(5):2629–2658, 2017. doi: 10.1109/TIT.2016.2642111.
- Jun S Liu. Metropolized independent sampling with comparisons to rejection sampling and importance sampling. Statistics and computing, 6(2):113–119, 1996.
- Monte Carlo strategies in scientific computing, volume 10. Springer, 2001.
- Vadim Lyubashevsky. Lattice signatures without trapdoors. In Annual International Conference on the Theory and Applications of Cryptographic Techniques, pages 738–755. Springer, 2012.
- A generalization of the adaptive rejection sampling algorithm. Statistics and Computing, 21(4):633–647, 2011.
- Reparameterization gradients through acceptance-rejection sampling algorithms. In Artificial Intelligence and Statistics, pages 489–498. PMLR, 2017.
- Quantum rejection sampling. ACM Transactions on Computation Theory (TOCT), 5(3):1–33, 2013.
- Y. Polyanskiy. Channel coding: non-asymptotic fundamental limits. PhD thesis, Princeton Univ., Princeton, NJ, USA, 2010. URL http://people.lids.mit.edu/yp/homepage.
- Information Theory: From Coding to Learning. Cambridge University Press, 2022+.
- Hans Rademacher. Über partielle und totale differenzierbarkeit von funktionen mehrerer variabeln und über die transformation der doppelintegrale. Mathematische Annalen, 79(4):340–359, 1919.
- Online learning: Stochastic, constrained, and smoothed adversaries. Advances in neural information processing systems, 24, 2011.
- Sequential complexities and uniform martingale laws of large numbers. Probability theory and related fields, 161(1):111–153, 2015.
- Relax and randomize: From value to algorithms. Advances in Neural Information Processing Systems, 25, 2012.
- Combinatorics of random processes and sections of convex bodies. Annals of Mathematics, pages 603–648, 2006.
- Rajan Srinivasan. Importance sampling: Applications in communications and detection. Springer Science & Business Media, 2002.
- Importance sampling: a review. Wiley Interdisciplinary Reviews: Computational Statistics, 2(1):54–60, 2010.
- Rényi divergence and kullback-leibler divergence. IEEE Transactions on Information Theory, 60(7):3797–3820, 2014.
- Ramon van Handel. The universal glivenko–cantelli property. Probability Theory and Related Fields, 155(3-4):911–934, 2013.
- Ramon Van Handel. Probability in high dimension. Technical report, PRINCETON UNIV NJ, 2014.
- John Von Neumann. 13. various techniques used in connection with random digits. Appl. Math Ser, 12(36-38):3, 1951.
- Martin J Wainwright. High-dimensional statistics: A non-asymptotic viewpoint, volume 48. Cambridge University Press, 2019.
- Rejection sampling revisit: How to choose parameters in lattice-based signature. Mathematical Problems in Engineering, 2021, 2021.