Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 134 tok/s
Gemini 2.5 Pro 41 tok/s Pro
GPT-5 Medium 27 tok/s Pro
GPT-5 High 24 tok/s Pro
GPT-4o 81 tok/s Pro
Kimi K2 205 tok/s Pro
GPT OSS 120B 432 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

Exploring Visual Prompting: Robustness Inheritance and Beyond (2506.06823v1)

Published 7 Jun 2025 in cs.CV and cs.AI

Abstract: Visual Prompting (VP), an efficient method for transfer learning, has shown its potential in vision tasks. However, previous works focus exclusively on VP from standard source models, it is still unknown how it performs under the scenario of a robust source model: Can the robustness of the source model be successfully inherited? Does VP also encounter the same trade-off between robustness and generalization ability as the source model during this process? If such a trade-off exists, is there a strategy specifically tailored to VP to mitigate this limitation? In this paper, we thoroughly explore these three questions for the first time and provide affirmative answers to them. To mitigate the trade-off faced by VP, we propose a strategy called Prompt Boundary Loosening (PBL). As a lightweight, plug-and-play strategy naturally compatible with VP, PBL effectively ensures the successful inheritance of robustness when the source model is a robust model, while significantly enhancing VP's generalization ability across various downstream datasets. Extensive experiments across various datasets show that our findings are universal and demonstrate the significant benefits of the proposed strategy.

Summary

  • The paper demonstrates that visual prompting inherits the robustness of adversarially trained models while exhibiting a trade-off with generalization accuracy.
  • It introduces the Prompt Boundary Loosening (PBL) strategy to expand decision boundaries and boost standard accuracy across various datasets.
  • Empirical evidence confirms that robust visual prompting refines feature representations, aligning with human perception in computer vision tasks.

Exploring Visual Prompting: Robustness Inheritance and Beyond

In the domain of computer vision, transferring knowledge efficiently from large-scale pre-trained models to target tasks is paramount. This paper explores Visual Prompting (VP), an innovative technique aimed at optimizing such transfer processes. With VP, the bulk parameters of a pre-trained model remain untouched, while only a minimal, learnable set of inputs is employed. This approach promises reduced computational cost and more seamless integration with diverse domains.

A significant aspect of this research is its exploration of VP in conjunction with robust source models. Robust models, often obtained through adversarial training, inherently stand resilient against adversarial attacks, yet they characteristically suffer from diminished accuracy on standard data. The research addresses how VP leverages the robustness of these models, raising pivotal inquiries: Can VP effectively inherit the robustness of robust models? Does VP encounter similar trade-offs between robustness and generalization as the models from which they derive?

Empirical evidence presented affirms that VP retains the robustness qualities intrinsic to robust models. Nevertheless, it similarly inherits the known trade-off, where robustness often compromises generalization performance. An introduction of the Prompt Boundary Loosening (PBL) strategy is set forth as a remedy. PBL operates as a lightweight, adaptable approach that seamlessly integrates with VP, broadening decision boundaries and thereby enhancing generalization without undermining robustness.

Quantitative validations underscore the universality and effectiveness of PBL across several datasets and robust source models. The experimental results elucidate significant boosts in standard accuracy through PBL, alongside sustained or improved adversarial robustness. Another notable assertion involves the intrinsic visual alignment of robust VP with human perception—a distinguishing trait absent in standard VP—as robust models refine feature representations fundamentally divergent from standard learning models.

This work provides a foundation for further investigation into optimizing transfer learning methodologies within adversarial contexts. The findings and strategies presented herein have significant implications for both theoretical advancements in model training paradigms and practical applications in robust AI systems. Moving forward, extending the scope of VP with adaptive strategies such as PBL might stimulate innovative solutions to perennial challenges in adversarial robust learning, providing scalability and reliability across sectors requiring high-stakes AI deployment.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Youtube Logo Streamline Icon: https://streamlinehq.com