MimiC: Combating Client Dropouts in Federated Learning by Mimicking Central Updates (2306.12212v4)
Abstract: Federated learning (FL) is a promising framework for privacy-preserving collaborative learning, where model training tasks are distributed to clients and only the model updates need to be collected at a server. However, when being deployed at mobile edge networks, clients may have unpredictable availability and drop out of the training process, which hinders the convergence of FL. This paper tackles such a critical challenge. Specifically, we first investigate the convergence of the classical FedAvg algorithm with arbitrary client dropouts. We find that with the common choice of a decaying learning rate, FedAvg oscillates around a stationary point of the global loss function, which is caused by the divergence between the aggregated and desired central update. Motivated by this new observation, we then design a novel training algorithm named MimiC, where the server modifies each received model update based on the previous ones. The proposed modification of the received model updates mimics the imaginary central update irrespective of dropout clients. The theoretical analysis of MimiC shows that divergence between the aggregated and central update diminishes with proper learning rates, leading to its convergence. Simulation results further demonstrate that MimiC maintains stable convergence performance and learns better models than the baseline methods.
- B. McMahan, E. Moore, D. Ramage, S. Hampson, and B. A. y Arcas, “Communication-efficient learning of deep networks from decentralized data,” in Proc. Int. Conf. Artif. Intell. Statist. (AISTATS), Ft. Lauderdale, FL, USA, Apr. 2017.
- K. Bonawitz et al., “Towards federated learning at scale: System design,” in Proc. Mach. Learn. Syst., Stanford, CA, USA, Mar. 2019.
- L. Li, Y. Fan, M. Tse, and K.-Y. Lin, “A review of applications in federated learning,” Comput. Ind. Eng., vol. 149, p. 106854, Nov. 2020.
- S. Wang, M. Chen, C. Yin, W. Saad, C. S. Hong, S. Cui, and H. V. Poor, “Federated learning for task and resource allocation in wireless high-altitude balloon networks,” IEEE Internet Things J., vol. 8, no. 24, pp. 17 460–17 475, Dec. 2021.
- H. Tong, Z. Yang, S. Wang, Y. Hu, W. Saad, and C. Yin, “Federated learning based audio semantic communication over wireless networks,” in Proc. IEEE Global Commun. Conf. (GLOBECOM), Madrid, Spain, Dec. 2021.
- W. Ni, J. Zheng, and H. Tian, “Semi-federated learning for collaborative intelligence in massive iot networks,” IEEE Internet Things J., vol. 10, no. 13, pp. 11 942 – 11 943, Jul. 2023.
- P. Kairouz et al., “Advances and open problems in federated learning,” Found. Trends Mach. Learn., vol. 14, no. 1–2, pp. 1–210, 2021.
- W. Y. B. Lim et al., “Federated learning in mobile edge networks: A comprehensive survey,” IEEE Commun. Surveys Tuts., vol. 22, no. 3, pp. 2031–2063, 3rd Quart. 2020.
- Z. Zhou, X. Chen, E. Li, L. Zeng, K. Luo, and J. Zhang, “Edge intelligence: Paving the last mile of artificial intelligence with edge computing,” Proc. IEEE, vol. 107, no. 8, pp. 1738–1762, Aug. 2019.
- G. Zhu, D. Liu, Y. Du, C. You, J. Zhang, and K. Huang, “Toward an intelligent edge: Wireless communication meets machine learning,” IEEE Commun. Mag., vol. 58, no. 1, pp. 19–25, Jan. 2020.
- C. Wang, X. Wei, and P. Zhou, “Optimize scheduling of federated learning on battery-powered mobile devices,” in Proc. IEEE Int. Parallel Distrib. Syst. Symp. (IPDPS), Virtual Event, May 2020.
- A. M. Abdelmoniem, C.-Y. Ho, P. Papageorgiou, and M. Canini, “Empirical analysis of federated learning in heterogeneous environments,” in Proc. 2nd Eur. Wkshop. Mach. Learn. Syst., Rennes, France, Apr. 2022.
- A. Imteaj, U. Thakker, S. Wang, J. Li, and M. H. Amini, “A survey on federated learning for resource-constrained IoT devices,” IEEE Internet Things J., vol. 9, no. 1, pp. 1–24, Jan. 2022.
- S. Li and S. Avestimehr, “Coded computing: Mitigating fundamental bottlenecks in large-scale distributed computing and machine learning,” Found. Trends Commun. Inf. Theory, vol. 17, no. 1, pp. 1–148, Aug. 2020.
- R. Bitar, M. Wootters, and S. El Rouayheb, “Stochastic gradient coding for straggler mitigation in distributed learning,” IEEE J. Sel. Areas Inf. Theory, vol. 1, no. 1, pp. 277–291, May 2020.
- G. Xiong, G. Yan, R. Singh, and J. Li, “Straggler-resilient distributed machine learning with dynamic backup workers,” [Online]. Available: https://arxiv.org/pdf/2102.06280.pdf.
- M. Ribero, H. Vikalo, and G. de Veciana, “Federated learning under intermittent client availability and time-varying communication constraints,” IEEE J. Sel. Topics Signal Process., vol. 17, no. 1, pp. 98–111, Jan. 2023.
- J. Wang, Q. Liu, H. Liang, G. Joshi, and H. V. Poor, “A novel framework for the analysis and design of heterogeneous federated learning,” IEEE Trans. Signal Process., vol. 69, pp. 5234–5249, 2021.
- H. Yang, M. Fang, and J. Liu, “Achieving linear speedup with partial worker participation in non-IID federated learning,” in Proc. Int. Conf. Learn. Repr. (ICLR), Virtual Event, May 2021.
- X. Li, K. Huang, W. Yang, S. Wang, and Z. Zhang, “On the convergence of fedavg on non-IID data,” [Online]. Available: https://arxiv.org/pdf/1907.02189.pdf.
- Y. Fraboni, R. Vidal, L. Kameni, and M. Lorenzi, “A general theory for client sampling in federated learning,” in Proc. Int. Wkshop. Trustworthy Federated Learning (FL-IJCAI’22), Vienna, Austria, Jul. 2022.
- Y. Yan, C. Niu, Y. Ding, Z. Zheng, F. Wu, G. Chen, S. Tang, and Z. Wu, “Distributed non-convex optimization with sublinear speedup under intermittent client availability,” [Online]. Available: https://arxiv.org/pdf/2002.07399.pdf.
- X. Gu, K. Huang, J. Zhang, and L. Huang, “Fast federated learning in the presence of arbitrary device unavailability,” in Proc. 35th Conf. Adv. Neural Inf. Process. Syst. (NeurIPS), Virtual Event, Dec. 2021.
- J. Xu and H. Wang, “Client selection and bandwidth allocation in wireless federated learning networks: A long-term perspective,” IEEE Trans. Wireless Commun., vol. 20, no. 2, pp. 1188–1200, Feb. 2021.
- Y. Ruan, X. Zhang, S.-C. Liang, and C. Joe-Wong, “Towards flexible device participation in federated learning,” in Proc Int. Conf. Artif. Intell. Statist. (AISTATS), Virtual Event, Apr. 2021.
- S. Wang and M. Ji, “A unified analysis of federated learning with arbitrary client participation,” in Proc. 35th Conf. Adv. Neural Inf. Process. Syst. (NeurIPS), LA, USA, Nov. 2022.
- F. Shi, C. Hu, W. Lin, L. Fan, T. Huang, and W. Wu, “VFedCS: Optimizing client selection for volatile federated learning,” IEEE Internet Things J., vol. 9, no. 24, pp. 24 995–25 010, Dec. 2022.
- C. Karakus, Y. Sun, S. Diggavi, and W. Yin, “Redundancy techniques for straggler mitigation in distributed optimization and learning,” J. Mach. Learn. Res., vol. 20, no. 72, pp. 1–47, 2019.
- S. Prakash et al., “Coded computing for low-latency federated learning over wireless edge networks,” IEEE J. Sel. Areas Commun., vol. 39, no. 1, pp. 233–250, Jan. 2020.
- Y. Sun, J. Shao, Y. Mao, S. Li, and J. Zhang, “Stochastic coded federated learning: Theoretical analysis and incentive mechanism design,” IEEE Trans. Wireless Commun., to appear.
- R. Schlegel, S. Kumar, E. Rosnes, and A. G. i Amat, “CodedPaddedFL and CodedSecAgg: Straggler mitigation and secure aggregation in federated learning,” IEEE Trans. Commun., vol. 71, no. 4, Apr. 2023.
- J. Shao, Y. Sun, S. Li, and J. Zhang, “DRes-FL: Dropout-resilient secure federated learning for non-IID clients via secret data sharing,” in Proc. 31st Conf. Adv. Neural Inf. Process. Syst. (NeurIPS), New Orleans, LA, USA, Nov. 2022.
- H. Wang and J. Xu, “Combating client dropout in federated learning via friend model substitution,” [Online]. Available: https://arxiv.org/pdf/2205.13222.pdf.
- Z. Jiang, W. Wang, and R. Chen, “Taming client dropout for distributed differential privacy in federated learning,” [Online]. Available: https://arxiv.org/pdf/2209.12528.pdf.
- Z. Liu, J. Guo, K.-Y. Lam, and J. Zhao, “Efficient dropout-resilient aggregation for privacy-preserving machine learning,” IEEE Trans. Inf. Forensics Secur., vol. 18, pp. 1839–1854, 2023.
- J. Wang, Q. Liu, H. Liang, G. Joshi, and H. V. Poor, “Tackling the objective inconsistency problem in heterogeneous federated optimization,” in Proc. 34th Conf. Adv. Neural Inf. Process. Syst. (NeurIPS), Virtual Event, Dec. 2020.
- T. Li, A. K. Sahu, M. Zaheer, M. Sanjabi, A. Talwalkar, and V. Smith, “Federated optimization in heterogeneous networks,” in Proc. Mach. Learn. Syst., Austin, TX, USA, Mar. 2020.
- S. P. Karimireddy, S. Kale, M. Mohri, S. Reddi, S. Stich, and A. T. Suresh, “SCAFFOLD: Stochastic controlled averaging for federated learning,” in Proc. Int. Conf. Mach. Learn. (ICML), Virtual Event, Jul. 2020.
- J. Perazzone, S. Wang, M. Ji, and K. S. Chan, “Communication-efficient device scheduling for federated learning using stochastic optimization,” in Proc. IEEE Int. Conf. Comput. Commun. (INFOCOM), Virtual Event, May 2022.
- S. U. Stich, “Local sgd converges fast and communicates little,” [Online]. Available: https://arxiv.org/pdf/1805.09767.pdf.
- F. Haddadpour, M. M. Kamani, M. Mahdavi, and V. Cadambe, “Local SGD with periodic averaging: Tighter analysis and adaptive synchronization,” in Proc. 33rd Conf. Adv. Neural Inf. Process. Syst. (NeurIPS), Vancouver, BC, Canada, Dec. 2019.
- A. Khaled, K. Mishchenko, and P. Richtárik, “Tighter theory for local SGD on identical and heterogeneous data,” in Proc. Int. Conf. Artif. Intell. Statist. (AISTATS), Virtual Event, Aug. 2020.
- L. Bottou, F. E. Curtis, and J. Nocedal, “Optimization methods for large-scale machine learning,” SIAM Rev., vol. 60, no. 2, pp. 223–311, Aug. 2018.
- A. Defazio, F. Bach, and S. Lacoste-Julien, “Saga: A fast incremental gradient method with support for non-strongly convex composite objectives,” in Proc. 28th Conf. Adv. Neural Inf. Process. Syst., Montréal, Canada, Dec. 2014.
- A. R. Elkordy and A. S. Avestimehr, “HeteroSAg: Secure aggregation with heterogeneous quantization in federated learning,” IEEE Trans. Commun., vol. 70, no. 4, pp. 2372–2386, Apr. 2022.
- K. Bonawitz et al., “Practical secure aggregation for privacy-preserving machine learning,” in Proc. ACM Conf. Comput. Commun. Secur. (CCS), Dallas, TX, USA, Oct. 2017.
- D. C. Nguyen, M. Ding, P. N. Pathirana, A. Seneviratne, J. Li, and H. V. Poor, “Federated learning for internet of things: A comprehensive survey,” IEEE Commun. Surveys Tuts., vol. 23, no. 3, pp. 1622–1658, 2021.
- H. Huang, R. Li, J. Liu, S. Zhou, K. Lin, and Z. Zheng, “ContextFL: Context-aware federated learning by estimating the training and reporting phases of mobile clients,” in Proc. Int. Conf. Distrib. Comput. Syst. (ICDCS), Bologna, Italy, Jul. 2022.
- H. Xiao, K. Rasul, and R. Vollgraf, “Fashion-MNIST: A novel image dataset for benchmarking machine learning algorithms,” [Online]. Available: https://arxiv.org/pdf/1708.07747.pdf.
- A. Krizhevsky et al., “Learning multiple layers of features from tiny images,” 2009.
- Q. Li, Y. Diao, Q. Chen, and B. He, “Federated learning on non-IID data silos: An experimental study,” [Online]. Available: https://arxiv.org/pdf/2102.02079.pdf.
- K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” [Online]. Available: https://arxiv.org/pdf/1409.1556.pdf, 2014.