Energy-Efficient Wireless Federated Learning via Doubly Adaptive Quantization (2402.12957v1)
Abstract: Federated learning (FL) has been recognized as a viable distributed learning paradigm for training a machine learning model across distributed clients without uploading raw data. However, FL in wireless networks still faces two major challenges, i.e., large communication overhead and high energy consumption, which are exacerbated by client heterogeneity in dataset sizes and wireless channels. While model quantization is effective for energy reduction, existing works ignore adapting quantization to heterogeneous clients and FL convergence. To address these challenges, this paper develops an energy optimization problem of jointly designing quantization levels, scheduling clients, allocating channels, and controlling computation frequencies (QCCF) in wireless FL. Specifically, we derive an upper bound identifying the influence of client scheduling and quantization errors on FL convergence. Under the longterm convergence constraints and wireless constraints, the problem is established and transformed into an instantaneous problem with Lyapunov optimization. Solving Karush-Kuhn-Tucker conditions, our closed-form solution indicates that the doubly adaptive quantization level rises with the training process and correlates negatively with dataset sizes. Experiment results validate our theoretical results, showing that QCCF consumes less energy with faster convergence compared with state-of-the-art baselines.
- B. McMahan, E. Moore, D. Ramage, S. Hampson, and B. A. y Arcas, “Communication-efficient learning of deep networks from decentralized data,” in Proc. Int. Conf. on Artif. Intell. Statist. PMLR, 2017, pp. 1273–1282.
- K. Wei, J. Li, M. Ding, C. Ma, H. H. Yang, F. Farokhi, S. Jin, T. Q. S. Quek, and H. V. Poor, “Federated learning with differential privacy: Algorithms and performance analysis,” IEEE Trans. Inf. Forensics Security, vol. 15, pp. 3454–3469, 2020.
- J. Li, Y. Shao, K. Wei, M. Ding, C. Ma, L. Shi, Z. Han, and H. V. Poor, “Blockchain assisted decentralized federated learning (BLADE-FL): Performance analysis and resource allocation,” IEEE Trans. Parallel Distrib. Syst., vol. 33, no. 10, pp. 2401–2415, 2022.
- S. Chu, J. Li, J. Wang, Z. Wang, M. Ding, Y. Zhang, Y. Qian, and W. Chen, “Federated learning over wireless channels: Dynamic resource allocation and task scheduling,” IEEE Trans. Cogn. Commun. Netw., vol. 8, no. 4, pp. 1910–1924, 2022.
- K. Fan, W. Chen, J. Li, X. Deng, X. Han, and M. Ding, “Mobility-aware joint user scheduling and resource allocation for low latency federated learning,” in Proc. IEEE/CIC Int. Conf. Commun. China, 2023, pp. 1–6.
- J. Wang, Q. Liu, H. Liang, G. Joshi, and H. V. Poor, “Tackling the objective inconsistency problem in heterogeneous federated optimization,” in Proc. Adv. Neural Inf. Process. Syst., vol. 33, 2020, pp. 7611–7623.
- Z. Jiang, Y. Xu, H. Xu, Z. Wang, J. Liu, Q. Chen, and C. Qiao, “Computation and communication efficient federated learning with adaptive model pruning,” IEEE Trans. Mobile Comput., pp. 1–18, 2023, early access.
- H. Q. Le, L. X. Nguyen, S.-B. Park, and C. S. Hong, “Layer-wise knowledge distillation for cross-device federated learning,” in Proc. IEEE Int. Conf. Inf. Netw., 2023, pp. 526–529.
- X. Deng, J. Li, C. Ma, K. Wei, L. Shi, M. Ding, and W. Chen, “Low-latency federated learning with DNN partition in distributed industrial IoT networks,” IEEE J. Sel. Areas Commun., vol. 41, no. 3, pp. 755–775, 2023.
- V. C, J. S, and R. B, “Incentive-based energy-efficient federated learning aggregation for intrusion detection in IoT sensor network,” in Proc. 14th Int. Conf. Comput. Commun. Netw. Technol., 2023, pp. 1–6.
- S. Han, C. Zhang, L. Wang, W. Zheng, and X. Wen, “Fedecs: Client selection for optimizing computing energy in federated learning,” in Proc. IEEE 33rd Annu. Int. Symp. Pers., Indoor Mobile Radio Commun., 2023, pp. 1–6.
- M. Alishahi, P. Fortier, W. Hao, X. Li, and M. Zeng, “Energy minimization for wireless-powered federated learning network with noma,” IEEE Wireless Commun. Lett., vol. 12, no. 5, pp. 833–837, 2023.
- R. Jin, X. He, and H. Dai, “Communication efficient federated learning with energy awareness over wireless networks,” IEEE Trans. Wireless Commun., vol. 21, no. 7, pp. 5204–5219, 2022.
- Y. Yang, Z. Zhang, and Q. Yang, “Communication-efficient federated learning with binary neural networks,” IEEE J. Sel. Areas Commun., vol. 39, no. 12, pp. 3836–3850, 2021.
- Y. Li, Y. Guo, M. Alazab, S. Chen, C. Shen, and K. Yu, “Joint optimal quantization and aggregation of federated learning scheme in vanets,” IEEE Trans. Intell. Transp. Syst., vol. 23, no. 10, pp. 19 852–19 863, 2022.
- S. Chen, C. Shen, L. Zhang, and Y. Tang, “Dynamic aggregation for heterogeneous quantization in federated learning,” IEEE Trans. Wireless Commun., vol. 20, no. 10, pp. 6804–6819, 2021.
- N. Shlezinger, M. Chen, Y. C. Eldar, H. V. Poor, and S. Cui, “UVeQFed: Universal vector quantization for federated learning,” IEEE Trans. Signal Process., vol. 69, pp. 500–514, 2021.
- P. Prakash, J. Ding, R. Chen, X. Qin, M. Shu, Q. Cui, Y. Guo, and M. Pan, “IoT device friendly and communication-efficient federated learning via joint model pruning and quantization,” IEEE Internet Things J., vol. 9, no. 15, pp. 13 638–13 650, 2022.
- Y.-J. Liu, G. Feng, D. Niyato, S. Qin, J. Zhou, X. Li, and X. Xu, “Ensemble distillation based adaptive quantization for supporting federated learning in wireless networks,” IEEE Trans. Wireless Commun., vol. 22, no. 6, pp. 4013–4027, 2023.
- D. Jhunjhunwala, A. Gadhikar, G. Joshi, and Y. C. Eldar, “Adaptive quantization of model updates for communication-efficient federated learning,” in Proc. IEEE Int. Conf. Acoust. Speech Signal Process., 2021, pp. 3110–3114.
- W. Feng and X. Zhang, “Wireless federated learning with dynamic quantization and bandwidth adaptation,” IEEE Wireless Commun. Lett., vol. 11, no. 11, pp. 2335–2339, 2022.
- A. Reisizadeh, A. Mokhtari, H. Hassani, A. Jadbabaie, and R. Pedarsani, “FedPAQ: A communication-efficient federated learning method with periodic averaging and quantization,” in Int. Conf. Artif. Intell. Statist. PMLR, 2020, pp. 2021–2031.
- Y. Mao, Z. Zhao, G. Yan, Y. Liu, T. Lan, L. Song, and W. Ding, “Communication-efficient federated learning with adaptive quantization,” ACM Trans. Intell. Syst. Technol., vol. 13, no. 4, pp. 1–26, 2022.
- R. Hönig, Y. Zhao, and R. Mullins, “DAdaQuant: Doubly-adaptive quantization for communication-efficient federated learning,” in Proc. 39th Int. Conf. on Mach. Learn., vol. 162. PMLR, 2022, pp. 8852–8866.
- A. M. Abdelmoniem and M. Canini, “Towards mitigating device heterogeneity in federated learning via adaptive model quantization,” in Proc. 1st Workshop Mach. Learn. Syst. Association for Computing Machinery, 2021, pp. 96––103.
- C. Battiloro, P. Di Lorenzo, M. Merluzzi, and S. Barbarossa, “Lyapunov-based optimization of edge resources for energy-efficient adaptive federated learning,” IEEE Trans. on Green Commun. Netw., vol. 7, no. 1, pp. 265–280, 2023.
- G. Chen, L. Yu, W. Luo, Y. Xu, and T. Song, “Rate-distortion optimization for adaptive gradient quantization in federated learning,” in Proc. IEEE Wireless Commun. Netw. Conf., 2023, pp. 1–6.
- M. M. Amiri, D. Gunduz, S. R. Kulkarni, and H. V. Poor, “Federated learning with quantized global model updates,” arXiv:2006.10672, 2020. [Online]. Available: https://arxiv.org/abs/2006.10672
- F. Wang, W. Chen, H. Tang, and Q. Wu, “Joint optimization of user association, subchannel allocation, and power allocation in multi-cell multi-association OFDMA heterogeneous networks,” IEEE Trans. Commun., vol. 65, no. 6, pp. 2672–2684, 2017.
- G. Wang, F. Gao, W. Chen, and C. Tellambura, “Channel estimation and training design for two-way relay networks in time-selective fading environments,” IEEE Trans. Wireless Commun., vol. 10, no. 8, pp. 2681–2691, 2011.
- R. Chen, L. Li, K. Xue, C. Zhang, M. Pan, and Y. Fang, “Energy efficient federated learning over heterogeneous mobile devices via joint design of weight quantization and wireless transmission,” IEEE Trans. Mobile Comput., pp. 1–13, 2022.
- 3GPP, “Study on channel model for frequencies from 0.5 to 100 GHz,” 3GPP TR 38.901, Tech. Rep., 2020.
- K. Wei, J. Li, C. Ma, M. Ding, C. Chen, S. Jin, Z. Han, and H. V. Poor, “Low-latency federated learning over wireless channels with differential privacy,” IEEE J. Sel. Areas Commun., vol. 40, no. 1, pp. 290–307, 2022.
- K. Tammer, “The application of parametric optimization and imbedding to the foundation and realization of a generalized primal decomposition approach,” Math. Res., vol. 35, pp. 376–386, 1987.
- A. Utami and Iskandar, “Optimization subcarrier allocation and genetic algorithm for resource allocation in MIMO-OFDMA,” in Proc. Int. Symp. Electron. Smart Devices, 2018, pp. 1–4.
- S. Caldas, P. Wu, T. Li, J. Konečný, H. B. McMahan, V. McMahan, and A. Talwalkar, “LEAF: A benchmark for federated settings,” arXiv:1812.01097, 2018. [Online]. Available: http://arxiv.org/abs/1812.01097
- A. Krizhevsky and G. Hinton, “Learning multiple layers of features from tiny images,” M.S. thesis, Dep. Computer Science, Univ. Toronto, Toronto, ON, Canada, Tech. Rep., 2009.
- Xuefeng Han (3 papers)
- Wen Chen (319 papers)
- Jun Li (780 papers)
- Ming Ding (219 papers)
- Qingqing Wu (263 papers)
- Kang Wei (41 papers)
- Xiumei Deng (6 papers)
- Zhen Mei (18 papers)