Understanding Threats to Federated Learning: A Survey
The paper "Threats to Federated Learning: A Survey" by Lingjuan Lyu, Han Yu, and Qiang Yang provides a comprehensive examination of security and privacy vulnerabilities in federated learning (FL) frameworks. As FL garners attention as a solution for decentralized data training with privacy considerations, understanding the complexity of potential attack vectors and their corresponding mitigation strategies becomes crucial. This survey bridges a vital gap by categorizing and analyzing various attacks on FL systems, thereby equipping future FL designers and researchers with insights to develop more robust systems.
Key Areas of Federated Learning Vulnerability
The paper identifies two primary categories of attacks on federated learning systems: poisoning attacks and inference attacks. Each type of attack exploits different aspects of the federated learning process and requires distinct defensive approaches.
- Poisoning Attacks: This category is further subdivided into data poisoning and model poisoning.
- In data poisoning, the adversary corrupts the training data to degrade the performance of the global model. Techniques include label-flipping and backdoor attacks.
- Model poisoning involves directly manipulating the model updates before submission to bias the global model's learning process without necessarily corrupting the input data.
- Inference Attacks: These attacks are focused on compromising the privacy of the data used in training.
- Attacks under this category include inferring class representatives, membership inference, property inference, and recovering training inputs and labels. The survey explores the powerful Deep Leakage from Gradient (DLG) attacks that can recover training samples from shared gradients.
FL Threat Models and Their Implications
The paper systematically explores the landscape of adversaries in federated settings, differentiating between insider vs. outsider threats and semi-honest vs. malicious adversaries. It highlights how these adversaries pose threats during both training and inference phases.
- Insider vs. Outsider: Insiders, which include compromised servers and participants, represent a more powerful threat than outsiders, as they can directly manipulate or observe the federated learning process.
- Semi-honest vs. Malicious: While semi-honest adversaries follow the protocol yet attempt to infer sensitive data, malicious ones actively deviate to degrade model integrity or privacy.
Defense Strategies and Research Directions
The survey assesses the current defense mechanisms, such as secure aggregation and differential privacy, stressing the limitations and application constraints, particularly in the context of different federated learning settings like horizontally federated learning (HFL) and vertically federated learning (VFL). There is a need to balance privacy, utility, and performance when deploying these measures.
Notably, the paper calls for more research in several areas:
- Developing federated learning protocols that minimize information leakage without compromising model performance.
- Exploring the robustness of federated models against attacks in VFL scenarios.
- Investigating federated learning with heterogeneous architectures and decentralized models.
- Enhancing the theoretical understanding of FL threats through interdisciplinary research, possibly using game-theoretic approaches to optimize defensive strategies.
In conclusion, the paper underscores emergent challenges in securing federated learning systems, advocating for rigorous threat assessments and adaptive defense mechanisms. The intricate balance of privacy, utility, and resilience remains a primary concern in the advancement of federated learning technologies, warranting a concerted effort from the research community to fortify emerging federated systems against evolving adversarial threats.