Papers
Topics
Authors
Recent
Search
2000 character limit reached

ScienceQA Benchmark for Multimodal Reasoning

Updated 1 February 2026
  • ScienceQA Benchmark is a large-scale multimodal dataset designed to evaluate machine reasoning and interpretability using chain-of-thought explanations.
  • It integrates diverse modalities—text, images, and combined formats—and is built on U.S. grade-school science curricula to challenge structured, multi-hop inference.
  • The benchmark drives innovation through modular architectures, retrieval-augmented prompting, and instruction tuning, achieving state-of-the-art performance.

ScienceQA Benchmark is a large-scale, multimodal science question answering (QA) benchmark widely adopted for the assessment and advancement of machine reasoning, interpretability, and multi-hop inference in natural and visual scientific contexts. Drawing from U.S. grade-school science curricula (grades 1–12), it sets a comprehensive standard for evaluating LLMs and multimodal models on reasoning tasks that require both factual knowledge and structured, stepwise explanation. Its structure has catalyzed a wave of methodological and architectural innovations, shaping the frontier of science-oriented AI research (Lu et al., 2022, Zhang et al., 2023, Horawalavithana et al., 2023, Lu et al., 2023, Liu et al., 2023).

1. Dataset Composition and Structure

ScienceQA comprises approximately 21,208 multiple-choice questions, each with a question stem, 2–5 answer options (avg. 4.4), and—depending on the item—a text passage, a scientific diagram or natural image, or no additional context. The dataset is factorially organized:

  • Subjects: Natural Science (NAT), Social Science (SOC), and Language Science (LAN), further subdivided into 26 topics, 127 categories, and 379 fine-grained skills.
  • Modalities:
    • Text-only (48.2%)
    • Image-only (48.7%, split into diagrams ~34.8% and natural photos ~14%)
    • Both image and text (30.8%)
    • No extra context (33.9%)
  • Grade Levels: Grades 1–2 (8.4%), 3–8 (72%), 9–12 (≈10%).

Annotation includes both “lecture” (background knowledge, ~83.9%) and “explanation” (multi-step chain-of-thought, ~90.5%) fields, enabling explicit supervision and analysis of multi-hop reasoning (Lu et al., 2022). Each explanation averages 47.7 tokens, and lectures 125.1 tokens.

The official data split is 60% train, 20% dev, 20% test, with curriculum and subject distributions balanced across partitions.

2. Evaluation Protocols and Metrics

ScienceQA employs strict accuracy as its primary performance metric,

Accuracy=number of correct predictionstotal questions\mathrm{Accuracy} = \frac{\text{number of correct predictions}}{\text{total questions}}

computed both overall and per subject, modality, and grade band. For generated explanations and rationales, automatic quality metrics (BLEU-n, ROUGE-L, sentence-level similarity via Sentence-BERT) and human evaluation along relevance, correctness, and completeness axes are used (Lu et al., 2022).

Prompting strategies vary from standard zero/few-shot QA to specialized chain-of-thought pipelines (prompting models to output answer → lecture → explanation), with a clear trend of improved stability and higher scores for multimodal and CoT-augmented approaches.

3. Methods and Architectural Innovations

3.1. Chain-of-Thought (CoT) Reasoning

Introduced in ScienceQA, explicit chain-of-thought supervision (ground-truth or model-generated) consistently improves performance in both few-shot (GPT-3: +1.2%) and fully supervised (UnifiedQA: +3.99%) regimes. Feeding models with provided lectures and explanations during inference produces an upper bound of 94% accuracy (GPT-3), quantifying the “explainability gap” relative to standard QA (Lu et al., 2022).

3.2. Modular and Compositional Reasoning

Chameleon generalizes the prior CoT pipeline into an LLM-based planner that orchestrates plug-and-play modules such as OCR, vision-language captioning, web search, and knowledge retrieval. The planner synthesizes module-execution plans (programs) tailored to each question. On ScienceQA, Chameleon with a GPT-4 planner achieves 86.54% few-shot accuracy, outperforming plain LLM invocation by >11 points. Ablations isolate module contributions (e.g., OCR: –8.4pp when disabled) and demonstrate reductions in solution and image-understanding error rates (Lu et al., 2023).

3.3. Multimodal CoT Reasoning

Multimodal-CoT, implemented as a two-stage encoder-decoder (vision + text; T5-Base/Large + ViT), first generates a multimodal rationale, then predicts the answer conditioned on it. Direct fusion via cross-attention and gated integration suppresses hallucination error rates by over 60% compared to text-only systems and drives state-of-the-art accuracy (T5-Large + ViT: 90.45%) with sub-1B parameter models (Zhang et al., 2023).

3.4. Multimodal Instruction Tuning

LLaMA-SciTune fuses CLIP-encoded visual representations with a LLaMA language decoder via pretrained lightweight adapters and is instruction-tuned on a dataset of 333k scientific figure–caption–paragraph tuples. Further finetuning on ScienceQA yields a 13B parameter model (CTOM) reaching 90.03% accuracy—surpassing human average (88.4%) and matching large-scale vision-language baselines without requiring proprietary inference backends (Horawalavithana et al., 2023). Additional modalities, e.g., OCR and text mentions, yield ∼1% accuracy improvement over captions alone.

3.5. Retrieval-Augmented Multimodal CoT

Retrieval-augmented chains-of-thought (CoT-MM-Retrieval) select demonstration exemplars tailored to each test question from the training set by cosine similarity in text/text, image/image, and cross-modal embedding spaces. Stratified sampling from these retrieved pools further boosts performance. This method lifts GPT-4’s ScienceQA accuracy from 86.5% (Chameleon) to 92.5% (k=4 demonstrations), a 6.0% absolute gain and current state-of-the-art. Ablations confirm that retrieved, visually similar examples are especially impactful for diagrammatic questions (Liu et al., 2023).

Approach ScienceQA Accuracy (%) Notable Base Model / Planner
Human (avg) 88.4
GPT-4 CoT 84.0 GPT-4
Chameleon 86.54 GPT-4 with modular tools
Multimodal-CoT 90.45 T5-Large + ViT
LLaMA-SciTune (13B CTOM) 90.03 LLaMA + CLIP
GPT-4 CoT-MM-Retrieval 92.5 GPT-4 + dynamic ex. retrieval
LLaVA+GPT-4 judge 92.53 LLaVA

4. Error Analysis and Challenges

Persistent model failures in ScienceQA predominantly arise from:

  • Commonsense gaps: Incorrect or incomplete “real-world” facts, especially in long reasoning chains (up to 94% of errors, LLaMA-SciTune) (Horawalavithana et al., 2023).
  • Logical inconsistencies: Contradictions within rationales (2–4%) (Horawalavithana et al., 2023, Zhang et al., 2023).
  • Image understanding errors: OCR failures, failure to parse diagrams or count elements, especially pronounced in image-based QA.
  • Partial or distractive CoTs: Rationale steps that are tangential or insufficient for answer determination.

Multimodal fusion and retrieval-augmented prompting mitigate (but do not eliminate) these error classes, with retrieval of visually similar demonstrations especially effective in diagrammatic tasks (Liu et al., 2023).

5. Extensions, Variants, and Generalization

The ScienceQA paradigm has inspired graduate-level variants that probe deeper domain and reasoning skills (e.g., MSQA for materials science), augmenting multiple-choice assessment with explanatory long answers and balanced binary judgments (Cheung et al., 29 May 2025). MSQA expands on ScienceQA by requiring models to synthesize self-contained, multi-step technical arguments, thus exposing failure modes like overfitting and distributional shift in domain-adapted LLMs and illuminating the utility of retrieval-augmentation and dual-format evaluation protocols.

6. Implications and Future Directions

ScienceQA’s layered annotation and multi-modal character have established a reproducible foundation for complex QA. Key lessons include:

  • Rich annotation (lectures and stepwise explanations) provides both a training signal for model reasoning and a diagnostic tool for error taxonomy.
  • Instruction tuning with human-written multi-modal prompts and early-fusion adapters can match or exceed “black-box” API accuracy.
  • Modular, planner-driven architectures (e.g., Chameleon) are extensible and can incorporate new modalities or external tools with minimal prompt engineering.
  • Retrieval-augmented prompting yields consistent and state-of-the-art gains by dynamically selecting in-context demonstrations based on cross-modal similarity.

A plausible implication is that future ScienceQA-style benchmarks in other domains (medicine, chemistry, engineering) will require stratified, multi-format evaluation, curated demonstration retrieval, and chain-of-thought supervision to faithfully measure and advance LLM reasoning (Liu et al., 2023, Cheung et al., 29 May 2025).

Topic to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to ScienceQA Benchmark.