Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Towards Adversarially Robust Vision-Language Models: Insights from Design Choices and Prompt Formatting Techniques (2407.11121v1)

Published 15 Jul 2024 in cs.CV, cs.AI, and cs.LG

Abstract: Vision-LLMs (VLMs) have witnessed a surge in both research and real-world applications. However, as they are becoming increasingly prevalent, ensuring their robustness against adversarial attacks is paramount. This work systematically investigates the impact of model design choices on the adversarial robustness of VLMs against image-based attacks. Additionally, we introduce novel, cost-effective approaches to enhance robustness through prompt formatting. By rephrasing questions and suggesting potential adversarial perturbations, we demonstrate substantial improvements in model robustness against strong image-based attacks such as Auto-PGD. Our findings provide important guidelines for developing more robust VLMs, particularly for deployment in safety-critical environments.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Rishika Bhagwatkar (3 papers)
  2. Shravan Nayak (11 papers)
  3. Reza Bayat (5 papers)
  4. Alexis Roger (5 papers)
  5. Pouya Bashivan (15 papers)
  6. Irina Rish (85 papers)
  7. Daniel Z Kaplan (1 paper)
X Twitter Logo Streamline Icon: https://streamlinehq.com