BlockDFL: A Blockchain-based Fully Decentralized Peer-to-Peer Federated Learning Framework (2205.10568v3)
Abstract: Federated learning (FL) enables collaborative training of machine learning models without sharing training data. Traditional FL heavily relies on a trusted centralized server. Although decentralized FL eliminates the central dependence, it may worsen the other inherit problems faced by FL such as poisoning attacks and data representation leakage due to insufficient restrictions on the behavior of participants, and heavy communication cost, especially in fully decentralized scenarios, i.e., peer-to-peer (P2P) settings. In this paper, we propose a blockchain-based fully decentralized P2P framework for FL, called BlockDFL. It takes blockchain as the foundation, leveraging the proposed PBFT-based voting mechanism and two-layer scoring mechanism to coordinate FL among peer participants without mutual trust, while effectively defending against poisoning attacks. Gradient compression is introduced to lowering communication cost and prevent data from being reconstructed from transmitted model updates. Extensive experiments conducted on two real-world datasets exhibit that BlockDFL obtains competitive accuracy compared to centralized FL and can defend poisoning attacks while achieving efficiency and scalability. Especially when the proportion of malicious participants is as high as 40%, BlockDFL can still preserve the accuracy of FL, outperforming existing fully decentralized P2P FL frameworks based on blockchain.
- Integration of blockchain and federated learning for Internet of Things: Recent advances and future challenges. Comput. Secur. 108 (2021), 102355.
- Poster: A Reliable and Accountable Privacy-Preserving Federated Learning Framework using the Blockchain. In ACM SIGSAC Conference on Computer and Communications Security. 2561–2563.
- Machine Learning with Adversaries: Byzantine Tolerant Gradient Descent. In Annual Conference on Neural Information Processing Systems. 119–129.
- Miguel Castro and Barbara Liskov. 1999. Practical Byzantine Fault Tolerance. In USENIX Symposium on Operating Systems Design and Implementation. 173–186.
- When Machine Learning Meets Blockchain: A Decentralized, Privacy-preserving and Secure Design. In IEEE International Conference on Big Data. 1178–1187.
- A Fast Blockchain-Based Federated Learning Framework With Compressed Communications. IEEE J. Sel. Areas Commun. 40, 12 (2022), 3358–3372.
- A Fast Blockchain-based Federated Learning Framework with Compressed Communications. IEEE Journal on Selected Areas in Communications 40, 12 (2022), 3358–3372.
- Li Deng. 2012. The mnist database of handwritten digit images for machine learning research. IEEE Signal Processing Magazine 29, 6 (2012), 141–142.
- BAFL: A Blockchain-Based Asynchronous Federated Learning Framework. IEEE Trans. Comput. 71, 5 (2022), 1092–1103.
- UCI machine learning repository, 2010. URL http://archive.ics.uci.edu/ml 15 (2011), 22.
- GPTQ: Accurate Post-Training Quantization for Generative Pre-trained Transformers. CoRR abs/2210.17323 (2022). https://doi.org/10.48550/arXiv.2210.17323 arXiv:2210.17323
- Model inversion attacks that exploit confidence information and basic countermeasures. In Proceedings of the 22nd ACM SIGSAC conference on computer and communications security. 1322–1333.
- Algorand: Scaling Byzantine Agreements for Cryptocurrencies. In the Symposium on Operating Systems Principles. 51–68.
- GFL: A Decentralized Federated Learning Framework Based On Blockchain. CoRR abs/2010.10996 (2020).
- Privacy-Preserving Decentralized Aggregation for Federated Learning. In IEEE Conference on Computer Communications Workshops. 1–6.
- An Efficient Blockchain Assisted Reputation Aware Decentralized Federated Learning Framework. IEEE Transactions on Network and Service Management 20, 3 (2023), 2771–2782. https://doi.org/10.1109/TNSM.2022.3231283
- Blockchained On-Device Federated Learning. IEEE Communications Letters 24, 6 (2020), 1279–1283.
- Learning multiple layers of features from tiny images. (2009).
- FedLGA: Toward System-Heterogeneity of Federated Learning via Local Gradient Approximation. IEEE Transactions on Cybernetics (2023), 1–14.
- Privacy-Preserving Federated Learning Framework Based on Chained Secure Multiparty Computing. IEEE Internet Things J. 8, 8 (2021), 6178–6186.
- Deep Gradient Compression: Reducing the Communication Bandwidth for Distributed Training. In International Conference on Learning Representations.
- Privacy-Enhanced Federated Learning Against Poisoning Adversaries. IEEE Transactions on Information Forensics and Security 16 (2021), 4574–4588.
- FedCoin: A Peer-to-Peer Payment System for Federated Learning. In Federated Learning - Privacy and Incentive. Lecture Notes in Computer Science, Vol. 12500. 125–138.
- Communication-Efficient Learning of Deep Networks from Decentralized Data. In the International Conference on Artificial Intelligence and Statistics, Vol. 54. 1273–1282.
- FedAPEN: Personalized Cross-silo Federated Learning with Adaptability to Statistical Heterogeneity. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 1954–1964.
- Biscotti: A Blockchain System for Private and Secure Federated Learning. IEEE Transactions on Parallel and Distributed Systems 32, 7 (2021), 1513–1525.
- Reza Shokri and Vitaly Shmatikov. 2015. Privacy-Preserving Deep Learning. In Proceedings of the Conference on Computer and Communications Security. 1310–1321.
- Soteria: Provable Defense Against Privacy Leakage in Federated Learning From Representation Perspective. In IEEE Conference on Computer Vision and Pattern Recognition. 9311–9319.
- Permissioned Blockchain Frame for Secure Federated Learning. IEEE Communications Letters 26, 1 (2022), 13–17. https://doi.org/10.1109/LCOMM.2021.3121297
- Laurens Van der Maaten and Geoffrey Hinton. 2008. Visualizing data using t-SNE. Journal of machine learning research 9, 11 (2008).
- FedEdge: Accelerating Edge-Assisted Federated Learning. In Proceedings of the ACM Web Conference 2023. 2895–2904.
- AI at the Edge: Blockchain-Empowered Secure Multiparty Learning With Heterogeneous Models. IEEE Internet of Things Journal 7, 10 (2020), 9600–9610.
- Against Membership Inference Attack: Pruning is All You Need. In the Joint Conference on Artificial Intelligence. 3141–3147.
- DeepChain: Auditable and Privacy-Preserving Deep Learning with Blockchain-Based Incentive. IEEE Transactions on Dependable and Secure Computing 18, 5 (2021), 2438–2455.
- Smoothquant: Accurate and efficient post-training quantization for large language models. In International Conference on Machine Learning. PMLR, 38087–38099.
- An accuracy-lossless perturbation method for defending privacy attacks in federated learning. In Proceedings of the ACM Web Conference 2022. 732–742.
- Communication-efficient and Scalable Decentralized Federated Edge Learning. In the Joint Conference on Artificial Intelligence. 5032–5035.
- Byzantine-robust distributed learning: Towards optimal statistical rates. In International Conference on Machine Learning. PMLR, 5650–5659.
- Incentive Mechanism for Horizontal Federated Learning Based on Reputation and Reverse Auction. In the Web Conference. 947–956.
- A Rate-and-Trust-Based Node Selection Model for Block Transmission in Blockchain Networks. IEEE Internet of Things Journal (2022).
- Peiyun Zhang and Mengchu Zhou. 2020. Security and trust in blockchains: Architecture, key technologies, and open issues. IEEE Transactions on Computational Social Systems 7, 3 (2020), 790–801.
- iDLG: Improved Deep Leakage from Gradients. CoRR abs/2001.02610 (2020).
- PVD-FL: A Privacy-Preserving and Verifiable Decentralized Federated Learning Framework. IEEE Transactions on Information Forensics and Security 17 (2022), 2059–2073.
- Privacy-Preserving Blockchain-Based Federated Learning for IoT Devices. IEEE Internet of Things Journal 8, 3 (2021), 1817–1829.
- A Blockchain-Based Approach for Saving and Tracking Differential-Privacy Cost. IEEE Internet Things J. 8, 11 (2021), 8865–8882.
- Blockchain-empowered Federated Learning: Challenges, Solutions, and Future Directions. ACM Comput. Surv. 55, 11 (2023), 240:1–240:31.
- Deep Leakage from Gradients. Advances in Neural Information Processing Systems 32 (2019), 14774–14784.
- Zhen Qin (105 papers)
- Xueqiang Yan (6 papers)
- Mengchu Zhou (32 papers)
- Shuiguang Deng (45 papers)