Adaptive Coded Federated Learning: Privacy Preservation and Straggler Mitigation (2403.14905v1)
Abstract: In this article, we address the problem of federated learning in the presence of stragglers. For this problem, a coded federated learning framework has been proposed, where the central server aggregates gradients received from the non-stragglers and gradient computed from a privacy-preservation global coded dataset to mitigate the negative impact of the stragglers. However, when aggregating these gradients, fixed weights are consistently applied across iterations, neglecting the generation process of the global coded dataset and the dynamic nature of the trained model over iterations. This oversight may result in diminished learning performance. To overcome this drawback, we propose a new method named adaptive coded federated learning (ACFL). In ACFL, before the training, each device uploads a coded local dataset with additive noise to the central server to generate a global coded dataset under privacy preservation requirements. During each iteration of the training, the central server aggregates the gradients received from the non-stragglers and the gradient computed from the global coded dataset, where an adaptive policy for varying the aggregation weights is designed. Under this policy, we optimize the performance in terms of privacy and learning, where the learning performance is analyzed through convergence analysis and the privacy performance is characterized via mutual information differential privacy. Finally, we perform simulations to demonstrate the superiority of ACFL compared with the non-adaptive methods.
- S. S. Band, S. Ardabili, M. Sookhak, A. T. Chronopoulos, S. Elnaffar, M. Moslehpour, M. Csaba, B. Torok, H.-T. Pai, and A. Mosavi, “When smart cities get smarter via machine learning: An in-depth literature review,” IEEE Access, vol. 10, pp. 60 985–61 015, 2022.
- S. V. Balkus, H. Wang, B. D. Cornet, C. Mahabal, H. Ngo, and H. Fang, “A survey of collaborative machine learning using 5g vehicular communications,” IEEE Communications Surveys & Tutorials, vol. 24, no. 2, pp. 1280–1303, 2022.
- K. H. Abdulkareem, M. A. Mohammed, A. Salim, M. Arif, O. Geman, D. Gupta, and A. Khanna, “Realizing an effective covid-19 diagnosis system based on machine learning and iot in smart hospital environment,” IEEE Internet of things journal, vol. 8, no. 21, pp. 15 919–15 928, 2021.
- P. Zhang, N. Chen, S. Li, K.-K. R. Choo, C. Jiang, and S. Wu, “Multi-domain virtual network embedding algorithm based on horizontal federated learning,” IEEE Transactions on Information Forensics and Security, 2023.
- L. U. Khan, W. Saad, Z. Han, E. Hossain, and C. S. Hong, “Federated learning for internet of things: Recent advances, taxonomy, and open challenges,” IEEE Communications Surveys & Tutorials, vol. 23, no. 3, pp. 1759–1799, 2021.
- M. Ye, X. Fang, B. Du, P. C. Yuen, and D. Tao, “Heterogeneous federated learning: State-of-the-art and research challenges,” ACM Computing Surveys, vol. 56, no. 3, pp. 1–44, 2023.
- H. Shao, C. Liu, X. Li, and D. Zhong, “Privacy preserving palmprint recognition via federated metric learning,” IEEE Transactions on Information Forensics and Security, 2023.
- B. McMahan, E. Moore, D. Ramage, S. Hampson, and B. A. y Arcas, “Communication-efficient learning of deep networks from decentralized data,” in Artificial intelligence and statistics. PMLR, 2017, pp. 1273–1282.
- C. Karakus, Y. Sun, S. Diggavi, and W. Yin, “Straggler mitigation in distributed optimization through data encoding,” Advances in Neural Information Processing Systems, vol. 30, 2017.
- M. Egger, M. Xhemrishi, A. Wachter-Zeh, and R. Bitar, “Sparse and private distributed matrix multiplication with straggler tolerance,” in 2023 IEEE International Symposium on Information Theory (ISIT). IEEE, 2023, pp. 2535–2540.
- S. Vahidian, M. Morafah, M. Shah, and B. Lin, “Rethinking data heterogeneity in federated learning: Introducing a new notion and standard benchmarks,” IEEE Transactions on Artificial Intelligence, 2023.
- A. M. Abdelmoniem, C.-Y. Ho, P. Papageorgiou, and M. Canini, “A comprehensive empirical study of heterogeneity in federated learning,” IEEE Internet of Things Journal, 2023.
- R. Tandon, Q. Lei, A. G. Dimakis, and N. Karampatziakis, “Gradient coding: Avoiding stragglers in distributed learning,” in International Conference on Machine Learning. PMLR, 2017, pp. 3368–3376.
- C. Li and M. Skoglund, “Distributed learning based on 1-bit gradient coding in the presence of stragglers,” accepted by IEEE Transactions on Communications, 2024.
- H. Wang, Z. Charles, and D. Papailiopoulos, “Erasurehead: Distributed gradient descent without delays using approximate gradient coding,” arXiv preprint arXiv:1901.09671, 2019.
- E. Ozfatura, D. Gündüz, and S. Ulukus, “Gradient coding with clustering and multi-message communication,” in 2019 IEEE Data Science Workshop (DSW). IEEE, 2019, pp. 42–46.
- B. Buyukates, E. Ozfatura, S. Ulukus, and D. Gündüz, “Gradient coding with dynamic clustering for straggler-tolerant distributed learning,” IEEE Transactions on Communications, 2022.
- M. Glasgow and M. Wootters, “Approximate gradient coding with optimal decoding,” IEEE Journal on Selected Areas in Information Theory, vol. 2, no. 3, pp. 855–866, 2021.
- R. Bitar, M. Wootters, and S. El Rouayheb, “Stochastic gradient coding for straggler mitigation in distributed learning,” IEEE Journal on Selected Areas in Information Theory, vol. 1, no. 1, pp. 277–291, 2020.
- Q. Li, Z. Wen, Z. Wu, S. Hu, N. Wang, Y. Li, X. Liu, and B. He, “A survey on federated learning systems: Vision, hype and reality for data privacy and protection,” IEEE Transactions on Knowledge and Data Engineering, 2021.
- S. Dhakal, S. Prakash, Y. Yona, S. Talwar, and N. Himayat, “Coded federated learning,” in 2019 IEEE Globecom Workshops (GC Wkshps). IEEE, 2019, pp. 1–6.
- S. Prakash, S. Dhakal, M. R. Akdeniz, Y. Yona, S. Talwar, S. Avestimehr, and N. Himayat, “Coded computing for low-latency federated learning over wireless edge networks,” IEEE Journal on Selected Areas in Communications, vol. 39, no. 1, pp. 233–250, 2020.
- A. Anand, S. Dhakal, M. Akdeniz, B. Edwards, and N. Himayat, “Differentially private coded federated linear regression,” in 2021 IEEE Data Science and Learning Workshop (DSLW). IEEE, 2021, pp. 1–6.
- Y. Sun, J. Shao, Y. Mao, S. Li, and J. Zhang, “Stochastic coded federated learning: Theoretical analysis and incentive mechanism design,” IEEE Transactions on Wireless Communications, 2023.
- M. Showkatbakhsh, C. Karakus, and S. Diggavi, “Privacy-utility trade-off of linear regression under random projections and additive noise,” in 2018 IEEE International Symposium on Information Theory (ISIT). IEEE, 2018, pp. 186–190.
- D. Maulud and A. M. Abdulazeez, “A review on linear regression comprehensive in machine learning,” Journal of Applied Science and Technology Trends, vol. 1, no. 4, pp. 140–147, 2020.
- J. Wen, B. Z. H. Zhao, M. Xue, A. Oprea, and H. Qian, “With great dispersion comes greater resilience: Efficient poisoning attacks and defenses for linear regression models,” IEEE Transactions on Information Forensics and Security, vol. 16, pp. 3709–3723, 2021.
- W. Liu, X. Mao, X. Zhang, and X. Zhang, “Efficient sparse least absolute deviation regression with differential privacy,” IEEE Transactions on Information Forensics and Security, 2024.
- A. Rahimi and B. Recht, “Random features for large-scale kernel machines,” Advances in neural information processing systems, vol. 20, 2007.
- P. Cuff and L. Yu, “Differential privacy as a mutual information constraint,” in Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, 2016, pp. 43–54.
- Q. Li, J. S. Gundersen, R. Heusdens, and M. G. Christensen, “Privacy-preserving distributed processing: metrics, bounds and algorithms,” IEEE Transactions on Information Forensics and Security, vol. 16, pp. 2090–2103, 2021.
- A. Rakhlin, O. Shamir, and K. Sridharan, “Making gradient descent optimal for strongly convex stochastic optimization,” arXiv preprint arXiv:1109.5647, 2011.
- K. Wei, J. Li, M. Ding, C. Ma, H. H. Yang, F. Farokhi, S. Jin, T. Q. Quek, and H. V. Poor, “Federated learning with differential privacy: Algorithms and performance analysis,” IEEE Transactions on Information Forensics and Security, vol. 15, pp. 3454–3469, 2020.
- K. Wei, J. Li, C. Ma, M. Ding, W. Chen, J. Wu, M. Tao, and H. V. Poor, “Personalized federated learning with differential privacy and convergence guarantee,” IEEE Transactions on Information Forensics and Security, 2023.
- X. Lyu, X. Hou, C. Ren, X. Ge, P. Yang, Q. Cui, and X. Tao, “Secure and efficient federated learning with provable performance guarantees via stochastic quantization,” IEEE Transactions on Information Forensics and Security, 2024.
- Chengxi Li (38 papers)
- Ming Xiao (178 papers)
- Mikael Skoglund (211 papers)