Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Vision-LLMs Can Fool Themselves with Self-Generated Typographic Attacks (2402.00626v2)

Published 1 Feb 2024 in cs.CV, cs.CR, and cs.LG

Abstract: Typographic Attacks, which involve pasting misleading text onto an image, were noted to harm the performance of Vision-LLMs like CLIP. However, the susceptibility of recent Large Vision-LLMs to these attacks remains understudied. Furthermore, prior work's Typographic attacks against CLIP randomly sample a misleading class from a predefined set of categories. However, this simple strategy misses more effective attacks that exploit LVLM(s) stronger language skills. To address these issues, we first introduce a benchmark for testing Typographic attacks against LVLM(s). Moreover, we introduce two novel and more effective \textit{Self-Generated} attacks which prompt the LVLM to generate an attack against itself: 1) Class Based Attack where the LVLM (e.g. LLaVA) is asked which deceiving class is most similar to the target class and 2) Descriptive Attacks where a more advanced LVLM (e.g. GPT4-V) is asked to recommend a Typographic attack that includes both a deceiving class and description. Using our benchmark, we uncover that Self-Generated attacks pose a significant threat, reducing LVLM(s) classification performance by up to 33\%. We also uncover that attacks generated by one model (e.g. GPT-4V or LLaVA) are effective against the model itself and other models like InstructBLIP and MiniGPT4. Code: \url{https://github.com/mqraitem/Self-Gen-Typo-Attack}

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Maan Qraitem (8 papers)
  2. Nazia Tasnim (9 papers)
  3. Kate Saenko (178 papers)
  4. Bryan A. Plummer (64 papers)
  5. Piotr Teterwak (16 papers)
Citations (6)

Summary

We haven't generated a summary for this paper yet.