KnFu: Effective Knowledge Fusion (2403.11892v1)
Abstract: Federated Learning (FL) has emerged as a prominent alternative to the traditional centralized learning approach. Generally speaking, FL is a decentralized approach that allows for collaborative training of Machine Learning (ML) models across multiple local nodes, ensuring data privacy and security while leveraging diverse datasets. Conventional FL, however, is susceptible to gradient inversion attacks, restrictively enforces a uniform architecture on local models, and suffers from model heterogeneity (model drift) due to non-IID local datasets. To mitigate some of these challenges, the new paradigm of Federated Knowledge Distillation (FKD) has emerged. FDK is developed based on the concept of Knowledge Distillation (KD), which involves extraction and transfer of a large and well-trained teacher model's knowledge to lightweight student models. FKD, however, still faces the model drift issue. Intuitively speaking, not all knowledge is universally beneficial due to the inherent diversity of data among local nodes. This calls for innovative mechanisms to evaluate the relevance and effectiveness of each client's knowledge for others, to prevent propagation of adverse knowledge. In this context, the paper proposes Effective Knowledge Fusion (KnFu) algorithm that evaluates knowledge of local models to only fuse semantic neighbors' effective knowledge for each client. The KnFu is a personalized effective knowledge fusion scheme for each client, that analyzes effectiveness of different local models' knowledge prior to the aggregation phase. Comprehensive experiments were performed on MNIST and CIFAR10 datasets illustrating effectiveness of the proposed KnFu in comparison to its state-of-the-art counterparts. A key conclusion of the work is that in scenarios with large and highly heterogeneous local datasets, local training could be preferable to knowledge fusion-based solutions.
- Konečnỳ, Jakub and McMahan, H Brendan and Yu, Felix X and Richtárik, Peter and Suresh, Ananda Theertha and Bacon, Dave, “Federated learning: Strategies for improving communication efficiency,” in arXiv preprint arXiv:1610.05492, 2016.
- Fu, Lei and Zhang, Huanle and Gao, Ge and Zhang, Mi and Liu, Xin, “Client selection in federated learning: Principles, challenges, and opportunities,” in IEEE Internet of Things Journal, 2023.
- Li, Daliang and Wang, Junpu, “Fedmd: Heterogenous federated learning via model distillation,” in arXiv preprint arXiv:1910.03581, 2019.
- Hinton, Geoffrey and Vinyals, Oriol and Dean, Jeff, “Distilling the knowledge in a neural network,” in arXiv preprint arXiv:1503.02531, 2015.
- S. Jamal Seyedmohammadi (4 papers)
- S. Kawa Atapour (2 papers)
- Jamshid Abouei (22 papers)
- Arash Mohammadi (69 papers)