Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

SA-Attack: Improving Adversarial Transferability of Vision-Language Pre-training Models via Self-Augmentation (2312.04913v1)

Published 8 Dec 2023 in cs.CV, cs.AI, cs.CR, and cs.LG

Abstract: Current Visual-Language Pre-training (VLP) models are vulnerable to adversarial examples. These adversarial examples present substantial security risks to VLP models, as they can leverage inherent weaknesses in the models, resulting in incorrect predictions. In contrast to white-box adversarial attacks, transfer attacks (where the adversary crafts adversarial examples on a white-box model to fool another black-box model) are more reflective of real-world scenarios, thus making them more meaningful for research. By summarizing and analyzing existing research, we identified two factors that can influence the efficacy of transfer attacks on VLP models: inter-modal interaction and data diversity. Based on these insights, we propose a self-augment-based transfer attack method, termed SA-Attack. Specifically, during the generation of adversarial images and adversarial texts, we apply different data augmentation methods to the image modality and text modality, respectively, with the aim of improving the adversarial transferability of the generated adversarial images and texts. Experiments conducted on the FLickr30K and COCO datasets have validated the effectiveness of our method. Our code will be available after this paper is accepted.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Bangyan He (2 papers)
  2. Xiaojun Jia (85 papers)
  3. Siyuan Liang (73 papers)
  4. Tianrui Lou (2 papers)
  5. Yang Liu (2253 papers)
  6. Xiaochun Cao (177 papers)
Citations (19)