Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

What Gives the Answer Away? Question Answering Bias Analysis on Video QA Datasets (2007.03626v1)

Published 7 Jul 2020 in cs.CL, cs.CV, cs.LG, and stat.ML

Abstract: Question answering biases in video QA datasets can mislead multimodal model to overfit to QA artifacts and jeopardize the model's ability to generalize. Understanding how strong these QA biases are and where they come from helps the community measure progress more accurately and provide researchers insights to debug their models. In this paper, we analyze QA biases in popular video question answering datasets and discover pretrained LLMs can answer 37-48% questions correctly without using any multimodal context information, far exceeding the 20% random guess baseline for 5-choose-1 multiple-choice questions. Our ablation study shows biases can come from annotators and type of questions. Specifically, annotators that have been seen during training are better predicted by the model and reasoning, abstract questions incur more biases than factual, direct questions. We also show empirically that using annotator-non-overlapping train-test splits can reduce QA biases for video QA datasets.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Jianing Yang (21 papers)
  2. Yuying Zhu (19 papers)
  3. Yongxin Wang (21 papers)
  4. Ruitao Yi (2 papers)
  5. Amir Zadeh (36 papers)
  6. Louis-Philippe Morency (123 papers)
Citations (12)

Summary

We haven't generated a summary for this paper yet.