Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
158 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Collaborative Active Learning in Conditional Trust Environment (2403.18436v1)

Published 27 Mar 2024 in cs.LG

Abstract: In this paper, we investigate collaborative active learning, a paradigm in which multiple collaborators explore a new domain by leveraging their combined machine learning capabilities without disclosing their existing data and models. Instead, the collaborators share prediction results from the new domain and newly acquired labels. This collaboration offers several advantages: (a) it addresses privacy and security concerns by eliminating the need for direct model and data disclosure; (b) it enables the use of different data sources and insights without direct data exchange; and (c) it promotes cost-effectiveness and resource efficiency through shared labeling costs. To realize these benefits, we introduce a collaborative active learning framework designed to fulfill the aforementioned objectives. We validate the effectiveness of the proposed framework through simulations. The results demonstrate that collaboration leads to higher AUC scores compared to independent efforts, highlighting the framework's ability to overcome the limitations of individual models. These findings support the use of collaborative approaches in active learning, emphasizing their potential to enhance outcomes through collective expertise and shared resources. Our work provides a foundation for further research on collaborative active learning and its practical applications in various domains where data privacy, cost efficiency, and model performance are critical considerations.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (18)
  1. G. Sachs, “Goldman sachs partners with apple on a game-changing credit card,” Goldman Sachs | Commemorates 150 Year History, 2019. [Online]. Available: https://www.goldmansachs.com/our-firm/history/moments/2019-apple-card.html
  2. Z. Kapron, “Malaysian digital banks: 29 applicants for a maximum of 5 licenses,” Forbes, Jul 2021. [Online]. Available: https://www.forbes.com/sites/zennonkapron/2021/07/19/malaysian-digital-banks-29-applicants-for-a-maximum-of-5-licenses/
  3. D. D. Lewis and J. Catlett, “Heterogeneous uncertainty sampling for supervised learning,” in Machine learning proceedings 1994.   Elsevier, 1994, pp. 148–156.
  4. Y. Freund, H. S. Seung, E. Shamir, and N. Tishby, “Selective sampling using the query by committee algorithm,” Machine learning, vol. 28, pp. 133–168, 1997.
  5. Z.-K. Chong, H. Ohsaki, and B.-M. Goi, “Improving uncertainty sampling with bell curve weight function,” International Journal of Applied Physics and Mathematics, vol. 13, no. 4, pp. 44–52, 2023.
  6. A. Tharwat and W. Schenck, “A survey on active learning: State-of-the-art, practical challenges and research directions,” Mathematics, vol. 11, no. 4, p. 820, 2023.
  7. Google. (2017) Federated learning: Collaborative machine learning without centralized training data. [Online]. Available: https://blog.research.google/2017/04/federated-learning-collaborative.html
  8. H. Wang, X. Liu, J. Niu, S. Tang, and J. Shen, “Unlocking the potential of federated learning for deeper models,” arXiv preprint arXiv:2306.02701, 2023.
  9. J. Wen, Z. Zhang, Y. Lan, Z. Cui, J. Cai, and W. Zhang, “A survey on federated learning: challenges and applications,” International Journal of Machine Learning and Cybernetics, vol. 14, no. 2, pp. 513–535, 2023.
  10. S. Warnat-Herresthal, H. Schultze, K. L. Shastry, S. Manamohan, S. Mukherjee, V. Garg, R. Sarveswara, K. Händler, P. Pickkers, N. A. Aziz et al., “Swarm learning for decentralized and confidential clinical machine learning,” Nature, vol. 594, no. 7862, pp. 265–270, 2021.
  11. J. Han, Y. Ma, and Y. Han, “Demystifying swarm learning: A new paradigm of blockchain-based decentralized federated learning,” arXiv preprint arXiv:2201.05286, 2022.
  12. O. L. Saldanha, P. Quirke, N. P. West, J. A. James, M. B. Loughrey, H. I. Grabsch, M. Salto-Tellez, E. Alwers, D. Cifci, N. Ghaffari Laleh et al., “Swarm learning for decentralized artificial intelligence in cancer histopathology,” Nature Medicine, vol. 28, no. 6, pp. 1232–1239, 2022.
  13. Z. Li, F. Mao, and C. Wu, “Can we share models if sharing data is not an option?” Patterns, vol. 3, no. 11, 2022.
  14. T. G. Dietterich, “Ensemble methods in machine learning,” in International workshop on multiple classifier systems.   Springer, 2000, pp. 1–15.
  15. T. T. Akano and C. C. James, “An assessment of ensemble learning approaches and single-based machine learning algorithms for the characterization of undersaturated oil viscosity,” Beni-Suef University Journal of Basic and Applied Sciences, vol. 11, no. 1, pp. 1–18, 2022.
  16. B. Naderalvojoud and T. Hernandez-Boussard, “Improving machine learning with ensemble learning on observational healthcare data,” in AMIA Annual Symposium Proceedings, vol. 2023.   American Medical Informatics Association, 2023, p. 521.
  17. Y. Li and W. Chen, “A comparative performance assessment of ensemble learning for credit scoring,” Mathematics, vol. 8, no. 10, p. 1756, 2020.
  18. F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay, “Scikit-learn: Machine learning in Python,” Journal of Machine Learning Research, vol. 12, pp. 2825–2830, 2011.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets