Papers
Topics
Authors
Recent
Search
2000 character limit reached

Diffusion Attack: Leveraging Stable Diffusion for Naturalistic Image Attacking

Published 21 Mar 2024 in cs.CV and eess.IV | (2403.14778v1)

Abstract: In Virtual Reality (VR), adversarial attack remains a significant security threat. Most deep learning-based methods for physical and digital adversarial attacks focus on enhancing attack performance by crafting adversarial examples that contain large printable distortions that are easy for human observers to identify. However, attackers rarely impose limitations on the naturalness and comfort of the appearance of the generated attack image, resulting in a noticeable and unnatural attack. To address this challenge, we propose a framework to incorporate style transfer to craft adversarial inputs of natural styles that exhibit minimal detectability and maximum natural appearance, while maintaining superior attack capabilities.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (7)
  1. Topiq: A top-down approach from semantics to distortions for image quality assessment. arXiv preprint arXiv:2308.03060, 2023.
  2. No-reference image quality assessment via transformers, relative ranking, and self-consistency. In Proceedings of the IEEE/CVF WACV, 2022.
  3. Q. Guo and J. Wen. Multi-level fusion based deep convolutional network for image quality assessment. In Pattern Recognition. ICPR International Workshops and Challenges: Virtual Event, January 10–15, 2021, Proceedings, Part VI, pp. 670–678. Springer, 2021.
  4. {{\{{SLAP}}\}}: Improving physical adversarial examples with {{\{{Short-Lived}}\}} adversarial perturbations. In 30th USENIX Security Symposium, 2021.
  5. A geometric convolutional neural network for 3d object detection. In 2019 IEEE Global Conference on Signal and Information Processing (GlobalSIP), pp. 1–5. IEEE, 2019.
  6. H. Talebi and P. Milanfar. Nima: Neural image assessment. IEEE TIP, 27(8):3998–4011, 2018.
  7. F. Woitschek and G. Schneider. Physical adversarial attacks on deep neural networks for traffic sign recognition: A feasibility study. In IEEE Intelligent Vehicles Symposium, 2021.

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.