Systematic Evaluation of Privacy Risks of Machine Learning Models
In this paper, the researchers present a methodical examination of the privacy risks inherent to machine learning models, particularly focusing on membership inference attacks. These attacks represent a significant privacy threat, as they are designed to determine whether a specific data point was part of the training set of a model—essentially breaching data privacy.
The authors critique existing work on this topic for relying heavily on neural network (NN)-based attacks to gauge these privacy risks, which can lead to an underestimation of the actual threat. They note that such methods often focus on aggregate attack accuracy metrics, which do not provide a complete picture of the risk landscape. Instead, they propose a comprehensive set of benchmark attacks that do not depend on NN predictors, many of which are based on refined calculations of prediction metrics such as prediction confidence and entropy.
Interestingly, this paper introduces a novel attack method that improves prediction entropy by considering the ground truth label, providing a more accurate measure of model susceptibility to membership inference. These benchmark attacks show that existing defenses, like adversarial regularization and MemGuard, are less effective than previously reported.
Further, the paper introduces a new metric, the privacy risk score, which assesses the risk of individual samples based on their probability of being training data. This allows for a more nuanced analysis, recognizing that privacy vulnerabilities can vary substantially between individual data points. The paper's experimental results highlight the heterogeneity in privacy risk scores, suggesting that average assessment methods may overlook high-risk individual samples.
The implications of this research extend to both the development of more robust privacy defenses and the theoretical understanding of model vulnerabilities. By employing a rigorous evaluation approach that considers fine-grained risks, better insight into machine learning privacy can be achieved. This work sets a precedence for future research, stressing the importance of systematic privacy risk evaluation.
Concluding, the research emphasizes the need for developing stronger defenses that account for adaptive adversaries and balance between privacy risk and model accuracy. The detailed experiments illustrate the intricacies of privacy risks, urging the research community to recognize and address these differences to advance secure and privacy-preserving machine learning architectures.