Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
167 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

On the Computational Landscape of Replicable Learning (2405.15599v2)

Published 24 May 2024 in cs.LG and stat.ML

Abstract: We study computational aspects of algorithmic replicability, a notion of stability introduced by Impagliazzo, Lei, Pitassi, and Sorrell [2022]. Motivated by a recent line of work that established strong statistical connections between replicability and other notions of learnability such as online learning, private learning, and SQ learning, we aim to understand better the computational connections between replicability and these learning paradigms. Our first result shows that there is a concept class that is efficiently replicably PAC learnable, but, under standard cryptographic assumptions, no efficient online learner exists for this class. Subsequently, we design an efficient replicable learner for PAC learning parities when the marginal distribution is far from uniform, making progress on a question posed by Impagliazzo et al. [2022]. To obtain this result, we design a replicable lifting framework inspired by Blanc, Lange, Malik, and Tan [2023] that transforms in a black-box manner efficient replicable PAC learners under the uniform marginal distribution over the Boolean hypercube to replicable PAC learners under any marginal distribution, with sample and time complexity that depends on a certain measure of the complexity of the distribution. Finally, we show that any pure DP learner can be transformed to a replicable one in time polynomial in the accuracy, confidence parameters and exponential in the representation dimension of the underlying hypothesis class.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (49)
  1. Private pac learning implies finite littlestone dimension. In Proceedings of the 51st Annual ACM SIGACT Symposium on Theory of Computing, pages 852–860, 2019.
  2. Private and online learnability are equivalent. ACM Journal of the ACM (JACM), 69(4):1–34, 2022.
  3. Monya Baker. 1,500 scientists lift the lid on reproducibility. Nature, 533(7604), 2016.
  4. Philip Ball. Is ai leading to a reproducibility crisis in science? Nature, 624(7990):22–25, 2023.
  5. Characterizing the sample complexity of private learners. In Proceedings of the 4th conference on Innovations in Theoretical Computer Science, pages 97–110, 2013.
  6. Popular decision tree algorithms are provably noise tolerant. In International Conference on Machine Learning, pages 2091–2106. PMLR, 2022a.
  7. Properly learning decision trees in almost polynomial time. Journal of the ACM, 69(6):1–19, 2022b.
  8. Lifting uniform learners via distributional decomposition. In Proceedings of the 55th Annual ACM Symposium on Theory of Computing, pages 1755–1767, 2023.
  9. Noise-tolerant learning, the parity problem, and the statistical query model. Journal of the ACM (JACM), 50(4):506–519, 2003.
  10. Practical privacy: the sulq framework. In Proceedings of the twenty-fourth ACM SIGMOD-SIGACT-SIGART symposium on Principles of database systems, pages 128–138, 2005.
  11. Avrim L Blum. Separating distribution-free and mistake-bound learning models over the boolean domain. SIAM Journal on Computing, 23(5):990–1000, 1994.
  12. Mark Bun. A computational separation between private learning and online learning. Advances in Neural Information Processing Systems, 33:20732–20743, 2020.
  13. An equivalence between private classification and online prediction. In 2020 IEEE 61st Annual Symposium on Foundations of Computer Science (FOCS), pages 389–402. IEEE, 2020.
  14. Stability is stable: Connections between replicability, privacy, and adaptive generalization. In Barna Saha and Rocco A. Servedio, editors, Proceedings of the 55th Annual ACM Symposium on Theory of Computing, STOC 2023, Orlando, FL, USA, June 20-23, 2023, pages 520–527. ACM, 2023. doi: 10.1145/3564246.3585246. URL https://doi.org/10.1145/3564246.3585246.
  15. Private PAC learning may be harder than online learning. In Claire Vernade and Daniel Hsu, editors, International Conference on Algorithmic Learning Theory, 25-28 February 2024, La Jolla, California, USA, volume 237 of Proceedings of Machine Learning Research, pages 362–389. PMLR, 2024. URL https://proceedings.mlr.press/v237/bun24a.html.
  16. Testing probability distributions using conditional samples. SIAM Journal on Computing, 44(3):540–616, 2015.
  17. Random restrictions of high dimensional distributions and uniformity testing with subcube conditioning. In Proceedings of the 2021 ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 321–336. SIAM, 2021.
  18. Local borsuk-ulam, stability, and replicability. arXiv preprint arXiv:2311.01599, 2023a.
  19. Stability and replicability in learning. In 64th IEEE Annual Symposium on Foundations of Computer Science, FOCS 2023, Santa Cruz, CA, USA, November 6-9, 2023, pages 2430–2439. IEEE, 2023b. doi: 10.1109/FOCS57990.2023.00148. URL https://doi.org/10.1109/FOCS57990.2023.00148.
  20. List and certificate complexities in replicable learning. In Alice Oh, Tristan Naumann, Amir Globerson, Kate Saenko, Moritz Hardt, and Sergey Levine, editors, Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023, New Orleans, LA, USA, December 10 - 16, 2023, 2023.
  21. Asymptotic minimax character of the sample distribution function and of the classical multinomial estimator. The Annals of Mathematical Statistics, pages 642–669, 1956.
  22. Calibrating noise to sensitivity in private data analysis. In Theory of Cryptography: Third Theory of Cryptography Conference, TCC 2006, New York, NY, USA, March 4-7, 2006. Proceedings 3, pages 265–284. Springer, 2006.
  23. Replicable reinforcement learning. In Alice Oh, Tristan Naumann, Amir Globerson, Kate Saenko, Moritz Hardt, and Sergey Levine, editors, Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023, New Orleans, LA, USA, December 10 - 16, 2023, 2023.
  24. Replicable bandits. In The Eleventh International Conference on Learning Representations, 2023a.
  25. Replicable clustering. In Alice Oh, Tristan Naumann, Amir Globerson, Kate Saenko, Moritz Hardt, and Sergey Levine, editors, Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023, New Orleans, LA, USA, December 10 - 16, 2023, 2023b.
  26. Statistical algorithms and a lower bound for detecting planted cliques. Journal of the ACM (JACM), 64(2):1–37, 2017.
  27. Efficient parameter estimation of truncated boolean product distributions. In Conference on Learning Theory, pages 1586–1600. PMLR, 2020.
  28. Efficient algorithms for learning from coarse labels. In Conference on Learning Theory, pages 2060–2079. PMLR, 2021.
  29. Perfect sampling from pairwise comparisons. In Sanmi Koyejo, S. Mohamed, A. Agarwal, Danielle Belgrave, K. Cho, and A. Oh, editors, Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, NeurIPS 2022, New Orleans, LA, USA, November 28 - December 9, 2022, 2022.
  30. Privacy induces robustness: Information-computation gaps and sparse mean estimation. Advances in Neural Information Processing Systems, 35:6829–6842, 2022.
  31. Sample-efficient proper pac learning with approximate differential privacy. In Proceedings of the 53rd Annual ACM SIGACT Symposium on Theory of Computing, pages 183–196, 2021a.
  32. User-level differentially private learning via correlated sampling. In Marc’Aurelio Ranzato, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, and Jennifer Wortman Vaughan, editors, Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, December 6-14, 2021, virtual, pages 20172–20184, 2021b.
  33. Statistical-query lower bounds via functional gradients. Advances in Neural Information Processing Systems, 33:2147–2158, 2020.
  34. How to construct random functions. Journal of the ACM (JACM), 33(4):792–807, 1986.
  35. Private learning implies online learning: An efficient reduction. Advances in Neural Information Processing Systems, 32, 2019.
  36. Faster sublinear algorithms using conditional sampling. In Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms, pages 1743–1757. SIAM, 2017.
  37. Privately releasing conjunctions and the statistical query barrier. In Proceedings of the forty-third annual ACM symposium on Theory of computing, pages 803–812, 2011.
  38. Reproducibility in learning. In Stefano Leonardi and Anupam Gupta, editors, STOC ’22: 54th Annual ACM SIGACT Symposium on Theory of Computing, Rome, Italy, June 20 - 24, 2022, pages 818–831. ACM, 2022. doi: 10.1145/3519935.3519973. URL https://doi.org/10.1145/3519935.3519973.
  39. Svante Janson. Tail bounds for sums of geometric and exponential variables. Statistics & Probability Letters, 135:1–6, 2018.
  40. Statistical indistinguishability of learning algorithms. In Andreas Krause, Emma Brunskill, Kyunghyun Cho, Barbara Engelhardt, Sivan Sabato, and Jonathan Scarlett, editors, International Conference on Machine Learning, ICML 2023, 23-29 July 2023, Honolulu, Hawaii, USA, volume 202 of Proceedings of Machine Learning Research, pages 15586–15622. PMLR, 2023. URL https://proceedings.mlr.press/v202/kalavasis23a.html.
  41. Replicable learning of large-margin halfspaces. arXiv preprint arXiv:2402.13857, 2024.
  42. Replicability in reinforcement learning. In Alice Oh, Tristan Naumann, Amir Globerson, Kate Saenko, Moritz Hardt, and Sergey Levine, editors, Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023, New Orleans, LA, USA, December 10 - 16, 2023, 2023.
  43. What can we learn privately? SIAM Journal on Computing, 40(3):793–826, 2011.
  44. Michael Kearns. Efficient noise-tolerant learning from statistical queries. Journal of the ACM (JACM), 45(6):983–1006, 1998.
  45. An introduction to computational learning theory. MIT press, 1994.
  46. Nick Littlestone. Learning quickly when irrelevant attributes abound: A new linear-threshold algorithm. Machine learning, 2:285–318, 1988.
  47. Pascal Massart. The tight constant in the dvoretzky-kiefer-wolfowitz inequality. The annals of Probability, pages 1269–1283, 1990.
  48. The bayesian stability zoo. In Alice Oh, Tristan Naumann, Amir Globerson, Kate Saenko, Moritz Hardt, and Sergey Levine, editors, Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023, New Orleans, LA, USA, December 10 - 16, 2023, 2023.
  49. Iclr reproducibility challenge 2019. ReScience C, 5(2):5, 2019.
Citations (1)

Summary

  • The paper establishes a computational separation by showing a concept class that is efficiently PAC learnable and replicable but not efficiently online learnable under standard cryptographic assumptions.
  • It introduces a replicable lifting framework that efficiently transforms learners from uniform to arbitrary marginal distributions using decision tree complexity.
  • It demonstrates that pure differentially private learners can be converted into replicable learners, albeit with exponential computational overhead relative to the concept class's representation dimension.

Computational Aspects of Algorithmic Replicability

Overview

This paper addresses the computational dimensions of algorithmic replicability, a stability notion introduced by Impagliazzo et al. The primary focus is to deepen our understanding of the computational connections between replicability and other learning paradigms, like online learning, private learning, and Statistical Query (SQ) learning. Significant theoretical and practical implications are drawn through both negative and positive results.

Key Contributions

Computational Separation from Online Learning

One major contribution of this work is to explore whether efficient replicability is computationally separable from efficient online learning. A pivotal result shows that under standard cryptographic assumptions, there exists a concept class that is efficiently PAC learnable and replicable but not efficiently learnable online.

Replicable Lifting Framework

The paper introduces an efficient framework for transforming replicable PAC learners under the uniform distribution to replicable PAC learners under any marginal distribution. The transformation's complexity depends on the decision tree complexity of the distribution. An essential implication here is the ability to learn parities efficiently and replicably under unknown distributions.

Transformation of Pure Differentially Private Learners

Another significant contribution is demonstrating that any pure differentially private (DP) learner can be transformed into a replicable learner. While the sample complexity remains polynomial in error, confidence, and replicability parameters, the computational complexity is exponential concerning the representation dimension of the concept class.

Detailed Analysis

Enhanced Understanding of Online Learning Separation

The computational separation between replicability and online learning is cemented through the concept class Σ\Sigma, constructed using pseudorandom function generators. This concept class exhibits the characteristic that, while efficiently PAC learnable and replicable, it resists efficient online learning under the assumption of one-way functions.

Efficient Replicable Learning of Parities

A robust replicable PAC learning process is designed for learning affine parities under non-uniform distributions. The transformation leverages decision tree complexity to ensure computational efficiency. The process includes partitioning the instance space based on a tree representation and using replicable learners for uniform marginals on these partitions.

The authors not only answer the posed question regarding learning affine parities under general distributions but extend their framework to demonstrate broader applicability.

Computational Transformations across Learning Paradigms

Attention is also given to the computational relationships between replicability and various stability notions such as DP and online learning. The authors confirm that there exists a concept class that is efficiently learnable under approximate DP but not under replicability, contingent on OWF assumptions. This result adds a nuanced layer to the overall understanding of the computational landscape of stable learning.

Computational Landscape of Stability

The results of this paper allow for a nuanced landscape of computational stability in learning frameworks to be drawn.

  • Black-Box Transformations: It is established that pure DP can be transformed to online learning; replicability can be transformed to approximate DP and so can SQ learning.
  • Separations: Notably, the paper delineates multiple separations:
    • Efficient learning through replicability but not online learning.
    • Approximate DP learning but not online learning.
    • Inefficient pure DP to replicable transformation compared to the more efficient approximate DP to replicable transformation under the same assumptions.

Future Directions and Implications

This paper sets a foundational ground for several future research directions.

  1. Complexity Measures: Further exploration is needed into alternative complexity measures beyond decision tree complexity. Understanding the trade-offs involved in these transformations can lead to more optimized algorithms.
  2. Extensions in Learning Paradigms: Extending the results to a broader range of concept classes and distributions would solidify these theoretical constructs further.
  3. Practical Implementations: While primarily theoretical, the results bear potential implications for practical applications, particularly in making machine learning experiments more replicable and reliable.
  4. Security Assumptions: The security assumptions (like OWF) underpin many results. Investigating weaker assumptions or different cryptographic bases can yield new insights.

Conclusion

By bridging gaps in understanding across replicability, online learning, SQ learning, and differential privacy, this paper sets an advanced discourse on the computational aspects of algorithmic replicability. The exploration not only allows for refined theoretical models but drives forward the practical dialogue on making machine learning outcomes more dependable in the real world.

X Twitter Logo Streamline Icon: https://streamlinehq.com