Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

DPL: Decoupled Prompt Learning for Vision-Language Models (2308.10061v1)

Published 19 Aug 2023 in cs.CV

Abstract: Prompt learning has emerged as an efficient and effective approach for transferring foundational Vision-LLMs (e.g., CLIP) to downstream tasks. However, current methods tend to overfit to seen categories, thereby limiting their generalization ability for unseen classes. In this paper, we propose a new method, Decoupled Prompt Learning (DPL), which reformulates the attention in prompt learning to alleviate this problem. Specifically, we theoretically investigate the collaborative process between prompts and instances (i.e., image patches/text tokens) by reformulating the original self-attention into four separate sub-processes. Through detailed analysis, we observe that certain sub-processes can be strengthened to bolster robustness and generalizability by some approximation techniques. Furthermore, we introduce language-conditioned textual prompting based on decoupled attention to naturally preserve the generalization of text input. Our approach is flexible for both visual and textual modalities, making it easily extendable to multi-modal prompt learning. By combining the proposed techniques, our approach achieves state-of-the-art performance on three representative benchmarks encompassing 15 image recognition datasets, while maintaining parameter-efficient. Moreover, our DPL does not rely on any auxiliary regularization task or extra training data, further demonstrating its remarkable generalization ability.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Chen Xu (186 papers)
  2. Yuhan Zhu (17 papers)
  3. Guozhen Zhang (14 papers)
  4. Haocheng Shen (7 papers)
  5. Yixuan Liao (6 papers)
  6. Xiaoxin Chen (25 papers)
  7. Gangshan Wu (70 papers)
  8. Limin Wang (221 papers)
Citations (4)