Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Survey: Leakage and Privacy at Inference Time (2107.01614v2)

Published 4 Jul 2021 in cs.LG

Abstract: Leakage of data from publicly available Machine Learning (ML) models is an area of growing significance as commercial and government applications of ML can draw on multiple sources of data, potentially including users' and clients' sensitive data. We provide a comprehensive survey of contemporary advances on several fronts, covering involuntary data leakage which is natural to ML models, potential malevolent leakage which is caused by privacy attacks, and currently available defence mechanisms. We focus on inference-time leakage, as the most likely scenario for publicly available models. We first discuss what leakage is in the context of different data, tasks, and model architectures. We then propose a taxonomy across involuntary and malevolent leakage, available defences, followed by the currently available assessment metrics and applications. We conclude with outstanding challenges and open questions, outlining some promising directions for future research.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Marija Jegorova (7 papers)
  2. Chaitanya Kaul (18 papers)
  3. Charlie Mayor (1 paper)
  4. Alison Q. O'Neil (23 papers)
  5. Alexander Weir (2 papers)
  6. Roderick Murray-Smith (42 papers)
  7. Sotirios A. Tsaftaris (100 papers)
Citations (58)

Summary

We haven't generated a summary for this paper yet.