Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Defending Against Weight-Poisoning Backdoor Attacks for Parameter-Efficient Fine-Tuning (2402.12168v3)

Published 19 Feb 2024 in cs.CR, cs.AI, and cs.CL

Abstract: Recently, various parameter-efficient fine-tuning (PEFT) strategies for application to LLMs have been proposed and successfully implemented. However, this raises the question of whether PEFT, which only updates a limited set of model parameters, constitutes security vulnerabilities when confronted with weight-poisoning backdoor attacks. In this study, we show that PEFT is more susceptible to weight-poisoning backdoor attacks compared to the full-parameter fine-tuning method, with pre-defined triggers remaining exploitable and pre-defined targets maintaining high confidence, even after fine-tuning. Motivated by this insight, we developed a Poisoned Sample Identification Module (PSIM) leveraging PEFT, which identifies poisoned samples through confidence, providing robust defense against weight-poisoning backdoor attacks. Specifically, we leverage PEFT to train the PSIM with randomly reset sample labels. During the inference process, extreme confidence serves as an indicator for poisoned samples, while others are clean. We conduct experiments on text classification tasks, five fine-tuning strategies, and three weight-poisoning backdoor attack methods. Experiments show near 100% success rates for weight-poisoning backdoor attacks when utilizing PEFT. Furthermore, our defensive approach exhibits overall competitive performance in mitigating weight-poisoning backdoor attacks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Shuai Zhao (116 papers)
  2. Leilei Gan (21 papers)
  3. Luu Anh Tuan (55 papers)
  4. Jie Fu (229 papers)
  5. Lingjuan Lyu (131 papers)
  6. Meihuizi Jia (5 papers)
  7. Jinming Wen (40 papers)
Citations (18)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com