Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
133 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Collaborative Learning with Different Labeling Functions (2402.10445v3)

Published 16 Feb 2024 in cs.LG, cs.DS, and stat.ML

Abstract: We study a variant of Collaborative PAC Learning, in which we aim to learn an accurate classifier for each of the $n$ data distributions, while minimizing the number of samples drawn from them in total. Unlike in the usual collaborative learning setup, it is not assumed that there exists a single classifier that is simultaneously accurate for all distributions. We show that, when the data distributions satisfy a weaker realizability assumption, which appeared in [Crammer and Mansour, 2012] in the context of multi-task learning, sample-efficient learning is still feasible. We give a learning algorithm based on Empirical Risk Minimization (ERM) on a natural augmentation of the hypothesis class, and the analysis relies on an upper bound on the VC dimension of this augmented class. In terms of the computational efficiency, we show that ERM on the augmented hypothesis class is NP-hard, which gives evidence against the existence of computationally efficient learners in general. On the positive side, for two special cases, we give learners that are both sample- and computationally-efficient.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (38)
  1. Neural Network Learning: Theoretical Foundations. Cambridge University Press, 1999.
  2. New approximation guarantee for chromatic number. In Symposium on Theory of Computing (STOC), pages 215–224, 2006.
  3. Open problem: The sample complexity of multi-distribution learning for vc classes. In Conference on Learning Theory (COLT), pages 5943–5949, 2023.
  4. A theory of learning from different domains. Machine learning, 79:151–175, 2010.
  5. Collaborative PAC learning. In Advances in Neural Information Processing Systems (NIPS), 2017.
  6. An Õ(n3/14superscript𝑛314n^{3/14}italic_n start_POSTSUPERSCRIPT 3 / 14 end_POSTSUPERSCRIPT)-coloring algorithm for 3-colorable graphs. Information Processing Letters, 61(1):49–53, 1997.
  7. Avrim Blum. New approximation algorithms for graph coloring. Journal of the ACM (JACM), 41(3):470–516, 1994.
  8. A better performance guarantee for approximate graph coloring. Algorithmica, 5(1-4):459–466, 1990.
  9. Federated asymptotics: a model to compare federated learning algorithms. In International Conference on Artificial Intelligence and Statistics (AISTATS), pages 10650–10689, 2023.
  10. Eden Chlamtac. Approximation algorithms using hierarchies of semidefinite programming relaxations. In Foundations of Computer Science (FOCS), pages 691–701, 2007.
  11. Minimax estimation for personalized federated learning: An alternative between fedavg and local training? Journal of Machine Learning Research, 24(262):1–59, 2023.
  12. Tight bounds for collaborative pac learning via multiplicative weights. In Advances in Neural Information Processing Systems (NeurIPS), pages 3602–3611, 2018.
  13. Efficient list-decodable regression using batches. In International Conference on Machine Learning (ICML), pages 7025–7065, 2023.
  14. Cryptographic hardness of learning halfspaces with massart noise. Advances in Neural Information Processing Systems (NeurIPS), 35:3624–3636, 2022.
  15. Near-optimal cryptographic hardness of agnostically learning halfspaces and relu regression under gaussian marginals. In International Conference on Machine Learning (ICML), pages 7922–7938, 2023.
  16. Vitaly Feldman. Optimal hardness results for maximizing agreements with monomials. In Conference on Computational Complexity (CCC), pages 226–236, 2006.
  17. New results for learning noisy parities and halfspaces. In Foundations of Computer Science (FOCS), pages 563–574, 2006.
  18. Hardness of learning halfspaces with noise. SIAM Journal on Computing, 39(2):742–765, 2009.
  19. On-demand sampling: Learning optimally from multiple distributions. In Advances in Neural Information Processing Systems (NeurIPS), pages 406–419, 2022.
  20. A no-free-lunch theorem for multitask learning. The Annals of Statistics, 50(6):3119–3143, 2022.
  21. Linear regression using heterogeneous data batches. arXiv preprint arXiv:2309.01973, 2023.
  22. Robust learning from untrusted sources. In International Conference on Machine Learning (ICML), pages 3488–3498, 2019.
  23. Approximate graph coloring by semidefinite programming. Journal of the ACM (JACM), 45(2):246–265, 1998.
  24. Robust meta-learning for mixed linear regression with small batches. In Advances in Neural Information Processing Systems (NeurIPS), volume 33, pages 4683–4696, 2020.
  25. Toward efficient agnostic learning. In Annual Workshop on Computational Learning Theory (COLT), pages 341–352, 1992.
  26. Meta-learning for mixed linear regression. In International Conference on Machine Learning (ICML), pages 5394–5404, 2020.
  27. Properly learning decision trees with queries is np-hard. In Foundations of Computer Science (FOCS), pages 2383–2407, 2023.
  28. Superpolynomial lower bounds for decision tree learning and testing. In Symposium on Discrete Algorithms (SODA), pages 1962–1994, 2023.
  29. Coloring 3-colorable graphs with less than n1/5 colors. Journal of the ACM (JACM), 64(1):1–23, 2017.
  30. Domain adaptation with multiple sources. In Advances in Neural Information Processing Systems (NIPS), 2008.
  31. Communication-efficient learning of deep networks from decentralized data. In International Conference on Artificial Intelligence and Statistics (AISTATS), pages 1273–1282, 2017.
  32. A theory of multiple-source adaptation with limited target labeled data. In International Conference on Artificial Intelligence and Statistics (AISTATS), pages 2332–2340, 2021.
  33. Agnostic federated learning. In International Conference on Machine Learning (ICML), pages 4615–4625, 2019.
  34. Improved algorithms for collaborative pac learning. In Advances in Neural Information Processing Systems (NeurIPS), volume 31, 2018.
  35. Binghui Peng. The sample complexity of multi-distribution learning. arXiv preprint arXiv:2312.04027, 2023.
  36. Understanding machine learning: From theory to algorithms. Cambridge university press, 2014.
  37. Avi Wigderson. Improving the performance guarantee for approximate graph coloring. Journal of the ACM (JACM), 30(4):729–735, 1983.
  38. Optimal multi-distribution learning. arXiv preprint arXiv:2312.05134, 2023.

Summary

We haven't generated a summary for this paper yet.