Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Towards Robust Visual Question Answering: Making the Most of Biased Samples via Contrastive Learning (2210.04563v1)

Published 10 Oct 2022 in cs.CV and cs.AI

Abstract: Models for Visual Question Answering (VQA) often rely on the spurious correlations, i.e., the language priors, that appear in the biased samples of training set, which make them brittle against the out-of-distribution (OOD) test data. Recent methods have achieved promising progress in overcoming this problem by reducing the impact of biased samples on model training. However, these models reveal a trade-off that the improvements on OOD data severely sacrifice the performance on the in-distribution (ID) data (which is dominated by the biased samples). Therefore, we propose a novel contrastive learning approach, MMBS, for building robust VQA models by Making the Most of Biased Samples. Specifically, we construct positive samples for contrastive learning by eliminating the information related to spurious correlation from the original training samples and explore several strategies to use the constructed positive samples for training. Instead of undermining the importance of biased samples in model training, our approach precisely exploits the biased samples for unbiased information that contributes to reasoning. The proposed method is compatible with various VQA backbones. We validate our contributions by achieving competitive performance on the OOD dataset VQA-CP v2 while preserving robust performance on the ID dataset VQA v2.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Qingyi Si (23 papers)
  2. Yuanxin Liu (28 papers)
  3. Fandong Meng (174 papers)
  4. Zheng Lin (104 papers)
  5. Peng Fu (43 papers)
  6. Yanan Cao (34 papers)
  7. Weiping Wang (123 papers)
  8. Jie Zhou (688 papers)
Citations (19)

Summary

We haven't generated a summary for this paper yet.