Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
87 tokens/sec
Gemini 2.5 Pro Premium
36 tokens/sec
GPT-5 Medium
31 tokens/sec
GPT-5 High Premium
39 tokens/sec
GPT-4o
95 tokens/sec
DeepSeek R1 via Azure Premium
91 tokens/sec
GPT OSS 120B via Groq Premium
460 tokens/sec
Kimi K2 via Groq Premium
219 tokens/sec
2000 character limit reached

Shortcuts Everywhere and Nowhere: Exploring Multi-Trigger Backdoor Attacks (2401.15295v3)

Published 27 Jan 2024 in cs.LG and cs.CR

Abstract: Backdoor attacks have become a significant threat to the pre-training and deployment of deep neural networks (DNNs). Although numerous methods for detecting and mitigating backdoor attacks have been proposed, most rely on identifying and eliminating the ``shortcut" created by the backdoor, which links a specific source class to a target class. However, these approaches can be easily circumvented by designing multiple backdoor triggers that create shortcuts everywhere and therefore nowhere specific. In this study, we explore the concept of Multi-Trigger Backdoor Attacks (MTBAs), where multiple adversaries leverage different types of triggers to poison the same dataset. By proposing and investigating three types of multi-trigger attacks including \textit{parallel}, \textit{sequential}, and \textit{hybrid} attacks, we demonstrate that 1) multiple triggers can coexist, overwrite, or cross-activate one another, and 2) MTBAs easily break the prevalent shortcut assumption underlying most existing backdoor detection/removal methods, rendering them ineffective. Given the security risk posed by MTBAs, we have created a multi-trigger backdoor poisoning dataset to facilitate future research on detecting and mitigating these attacks, and we also discuss potential defense strategies against MTBAs. Our code is available at \url{https://github.com/bboylyg/Multi-Trigger-Backdoor-Attacks}.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (43)
  1. A new backdoor attack in cnns by training set corruption without label poisoning. In ICIP, 2019.
  2. Coyo-700m: Image-text pair dataset. https://github.com/kakaobrain/coyo-dataset, 2022.
  3. Poisoning web-scale training datasets is practical. arXiv preprint arXiv:2302.10149, 2023.
  4. Listen, attend and spell: A neural network for large vocabulary conversational speech recognition. In ICASSP, 2016.
  5. Detecting backdoor attacks on deep neural networks by activation clustering. In AAAI Workshop, 2019.
  6. Targeted backdoor attacks on deep learning systems using data poisoning. arXiv preprint arXiv:1712.05526, 2017.
  7. Deep feature space trojan attack of neural networks by controlled detoxification. In AAAI, 2021.
  8. Imagenet: A large-scale hierarchical image database. In CVPR, 2009.
  9. Bert: Pre-training of deep bidirectional transformers for language understanding. In NAACL, 2019.
  10. An image is worth 16x16 words: Transformers for image recognition at scale. 2021.
  11. Strip: A defence against trojan attacks on deep neural networks. In ACSAC, 2019.
  12. Badnets: Identifying vulnerabilities in the machine learning model supply chain. arXiv preprint arXiv:1708.06733, 2017.
  13. Tabor: A highly accurate approach to inspecting and restoring trojan backdoors in ai systems. arXiv preprint arXiv:1908.01763, 2019.
  14. Deep residual learning for image recognition. In CVPR, 2016.
  15. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861, 2017.
  16. Distilling cognitive backdoor patterns within an image. ICLR, 2023.
  17. Backdoor defense via decoupling the training process. In ICLR, 2022.
  18. Rethinking the trigger of backdoor attack. arXiv preprint arXiv:2004.04692, 2020.
  19. Invisible backdoor attack with sample-specific triggers. In ICCV, pp.  16463–16472, 2021a.
  20. Anti-backdoor learning: Training clean models on poisoned data. In NeurIPS, 2021b.
  21. Neural attention distillation: Erasing backdoor triggers from deep neural networks. In ICLR, 2021c.
  22. Backdoor learning: A survey. TNNLS, 2022.
  23. Reconstructive neuron pruning for backdoor defense. In ICML, 2023.
  24. Composite backdoor attack for deep neural network by mixing existing benign features. In CCS, 2020.
  25. Fine-pruning: Defending against backdooring attacks on deep neural networks. In RAID, 2018a.
  26. Trojaning attack on neural networks. In NDSS, 2018b.
  27. Abs: Scanning neural networks for back-doors by artificial brain stimulation. In Proceedings of the 2019 ACM SIGSAC Conference on Computer and Communications Security, pp.  1265–1282, 2019.
  28. Reflection backdoor: A natural backdoor attack on deep neural networks. In ECCV, 2020.
  29. Language models are few-shot learners. 2020.
  30. Input-aware dynamic backdoor attack. In NeurIPS, 2020.
  31. Wanet–imperceptible warping-based backdoor attack. In ICLR, 2021.
  32. Laion-400m: Open dataset of clip-filtered 400 million image-text pairs. arXiv preprint arXiv:2111.02114, 2021.
  33. Spectral signatures in backdoor attacks. In NeurIPS, 2018.
  34. Clean-label backdoor attacks. https://people.csail.mit.edu/madry/lab/, 2019.
  35. Visualizing data using t-sne. Journal of machine learning research, 9(11), 2008.
  36. Neural cleanse: Identifying and mitigating backdoor attacks in neural networks. In S&P. IEEE, 2019.
  37. Mm-bd: Post-training detection of backdoor attacks with arbitrary backdoor pattern types using a maximum margin statistic. In Symposium on Security and Privacy (SP), 2023.
  38. Adversarial neuron pruning purifies backdoored deep models. NeurIPS, 2021.
  39. Umd: Unsupervised model detection for x2x backdoor attacks. In ICML, 2023.
  40. Detecting ai trojans using meta neural analysis. In S&P, 2021.
  41. One-to-n & n-to-one: Two advanced backdoor attacks against deep learning models. TDSC, 2020.
  42. Rethinking the backdoor attacks’ triggers: A frequency perspective. In ICCV, 2021.
  43. Data-free backdoor removal based on channel lipschitzness. In ECCV, 2022.
Citations (5)
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.