Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Threats to Pre-trained Language Models: Survey and Taxonomy (2202.06862v1)

Published 14 Feb 2022 in cs.CR

Abstract: Pre-trained LLMs (PTLMs) have achieved great success and remarkable performance over a wide range of NLP tasks. However, there are also growing concerns regarding the potential security issues in the adoption of PTLMs. In this survey, we comprehensively systematize recently discovered threats to PTLM systems and applications. We perform our attack characterization from three interesting perspectives. (1) We show threats can occur at different stages of the PTLM pipeline raised by different malicious entities. (2) We identify two types of model transferability (landscape, portrait) that facilitate attacks. (3) Based on the attack goals, we summarize four categories of attacks (backdoor, evasion, data privacy and model privacy). We also discuss some open problems and research directions. We believe our survey and taxonomy will inspire future studies towards secure and privacy-preserving PTLMs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Shangwei Guo (32 papers)
  2. Chunlong Xie (3 papers)
  3. Jiwei Li (137 papers)
  4. Lingjuan Lyu (131 papers)
  5. Tianwei Zhang (199 papers)
Citations (29)