The Illusion of Readiness: Stress Testing Large Frontier Models on Multimodal Medical Benchmarks (2509.18234v1)
Abstract: Large frontier models like GPT-5 now achieve top scores on medical benchmarks. But our stress tests tell a different story. Leading systems often guess correctly even when key inputs like images are removed, flip answers under trivial prompt changes, and fabricate convincing yet flawed reasoning. These aren't glitches; they expose how today's benchmarks reward test-taking tricks over medical understanding. We evaluate six flagship models across six widely used benchmarks and find that high leaderboard scores hide brittleness and shortcut learning. Through clinician-guided rubric evaluation, we show that benchmarks vary widely in what they truly measure yet are treated interchangeably, masking failure modes. We caution that medical benchmark scores do not directly reflect real-world readiness. If we want AI to earn trust in healthcare, we must demand more than leaderboard wins and must hold systems accountable for robustness, sound reasoning, and alignment with real medical demands.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Explain it Like I'm 14
Overview
This paper looks at whether super-powerful AI models (like GPT-5) are truly ready to help with medical tasks that involve both text and images. Even though these models get top scores on popular medical tests, the authors show that high scores can be misleading. The models often succeed by using test-taking tricks instead of real medical understanding.
What questions does the paper ask?
The paper asks simple but important questions:
- Are big AI models answering medical questions for the right reasons?
- Do they actually understand medical images and text together, or are they guessing based on patterns?
- Are current medical benchmarks (the tests we use to measure AI) really checking what matters for real healthcare?
- How can we better test models to make sure they’re reliable, safe, and trustworthy?
How did the researchers test this?
The team ran “stress tests” on six leading models across six common medical benchmarks. Think of stress tests like shaking a table to see if it’s sturdy or just looks good.
They focused on multimodal tasks, which means questions that combine words and pictures (like a short patient story plus a chest X-ray or a skin photo). Here’s what they did, in everyday terms:
- Removed images to see if models still answer “image-required” questions correctly. If they do, they may be guessing from text patterns instead of understanding visuals.
- Shuffled answer choices to check whether models rely on answer position (like “C is usually right”) rather than content.
- Replaced wrong answer options with random alternatives, or with “Unknown,” to see if models handle uncertainty or just do elimination tricks.
- Swapped the original image with one that visually matches a different answer to test whether models truly integrate visuals and text.
- Asked models to explain their reasoning (“think step by step”) and checked whether the explanations were factual and grounded in the images.
They also created a “robustness score,” a number that summarizes how stable a model is when inputs are incomplete or tricky. Finally, they asked clinicians to rate several benchmarks on how much reasoning and visual detail each one really needs. This helps reveal what those tests are actually measuring.
What did they find?
The results show hidden weaknesses that leaderboard scores don’t capture:
- Models succeed without the image: On questions that require pictures, most models still scored way above random even after the image was removed. That suggests they rely on shortcuts—like common word patterns or memorized question-answer pairs—rather than true visual understanding.
- Brittle under small changes: Simply reordering answer options lowered accuracy in text-only settings, which means the models were sensitive to format rather than content. Replacing distractors (wrong options) with random ones pushed text-only performance down toward guessing.
- Visual swap trap: When the image was replaced with one that supported a different answer, accuracy dropped sharply—even though the text didn’t change. This shows models often match images to labels superficially instead of reasoning with both text and visuals together.
- Reasoning often looks good but is wrong: When asked to explain their answers, models could produce confident, detailed medical-sounding explanations—but often based on features not actually in the image or on flawed logic. In many cases, asking for step-by-step reasoning didn’t improve accuracy.
- Benchmarks measure different things: Clinician ratings showed that popular datasets vary widely. For example, NEJM questions tend to require both deep reasoning and careful image reading, while JAMA can often be answered from text alone. Other datasets may be heavy on image detection but light on inference. Yet these tests are often treated as if they measure the same skill.
Why this matters: In real medicine, information is messy, uncertain, and high-stakes. If an AI gets confused when answer choices are shuffled or when a picture changes, it’s risky to rely on it for diagnosis or treatment decisions.
What does this mean for the future of medical AI?
The paper urges the field to rethink how progress is measured and to raise the bar for readiness:
- Don’t trust leaderboard scores alone. High accuracy doesn’t guarantee robust understanding, safe behavior under uncertainty, or sound medical reasoning.
- Treat benchmarks like diagnostic tools. Report performance along meaningful dimensions (like visual complexity and reasoning depth) instead of averaging across very different tasks.
- Include stress tests in evaluations. Regularly test models with missing images, format changes, tricky distractors, and reasoning audits—then report those results.
- Share benchmark profiles. Datasets should include clinician-informed metadata describing what they test (e.g., image sensitivity, reasoning complexity).
- Align evaluation with real-world needs. Pick tests that match the intended clinical use, and demand evidence of reliability, caution when information is missing, and clear, truthful reasoning.
In short: If we want AI to earn trust in healthcare, we must make sure it performs well for the right reasons—not just because it can pass a test.
Knowledge Gaps
Knowledge gaps, limitations, and open questions
Below is a concise, actionable list of what remains missing, uncertain, or unexplored based on the paper’s methods, analyses, and claims.
- Quantifying data contamination: No measurement of training–test overlap or memorization for closed-source models; need de-duplication audits and leakage checks for each benchmark item.
- Attribution of above-chance text-only performance: Unclear contributions of priors vs. co-occurrence artifacts vs. memorized QA pairs; requires controlled synthetic items and counterfactuals to disentangle causes.
- Limited modality scope: Focus on image+text; no evaluation of other clinical modalities (e.g., time-series vitals, waveforms, EHR notes, lab trajectories, pathology whole-slide, ultrasound cine, genomics).
- Single-snapshot tasks: Absence of tasks requiring longitudinal/temporal reasoning, prior comparisons, or multi-view alignment that mirror real clinical workflows.
- Benchmark coverage and representativeness: Evaluation centered on a small set of popular datasets; unclear generalizability to specialty domains (e.g., pediatrics, oncology, ophthalmology beyond fundus) and uncommon conditions.
- Small, curated perturbation samples: Stress tests rely on limited subsets (e.g., 175 NEJM items; 40 visual substitutions), raising questions about statistical power and representativeness; need larger, stratified, and preregistered perturbation sets.
- Statistical uncertainty: No confidence intervals, bootstrap analyses, or tests of significance across runs and prompts; robustness findings may be sensitive to stochasticity.
- Run-to-run and prompt sensitivity: Limited analysis of temperature, decoding strategy, system prompts, and instruction formatting; needs systematic sensitivity sweeps and variance quantification.
- Faithfulness of perturbations: Visual substitution may alter difficulty or introduce unintended cues; requires validated counterfactual generation protocols with clinician sign-off and item-level difficulty calibration.
- External validity to real settings: No prospective or simulated-user studies to link robustness scores to clinician performance, workflow impact, or patient safety outcomes.
- Lack of abstention-aware evaluation: Accuracy penalizes refusal; no metrics for appropriate abstention, selective prediction, or utility under uncertainty (e.g., coverage vs. risk, decision-curve analysis).
- Calibration gaps: No assessment of probability calibration, error severity weighting, or cost-sensitive metrics; need reliability diagrams and clinically weighted loss functions.
- Safety-critical error taxonomy: Absence of stratified error analysis by clinical risk (benign vs. harmful mistakes); actionable safety benchmarks remain undefined.
- Reasoning faithfulness measurement: Manual audits reveal fabricated rationales, but no standardized, automated metrics for visual grounding and explanation faithfulness in multimodal settings.
- Intervention studies missing: Paper diagnoses shortcuts but does not test training or decoding interventions (e.g., adversarial fine-tuning, uncertainty-aware objectives, debiasing, evidence-grounded rationales) that could reduce brittleness.
- Benchmark metadata granularity: Rubric applied at benchmark level, not item level; item-wise annotations (e.g., visual necessity, reasoning depth) are needed for stratified evaluation and dataset repair.
- Clinician rubric validity: Only three annotators per benchmark and moderate agreement on subjective axes; requires broader, multi-institution validation and outcome-linked construct validity.
- Standardization of stress testing: No community reference suite or reproducible harness with canonical perturbation generators, acceptance criteria, and pass/fail thresholds.
- Aggregation choices in robustness score: Weighting across tests is not justified; no analysis of sensitivity to weights or correlation of robustness score with downstream safety/utility.
- Closed-model opacity: Heavy reliance on proprietary models limits reproducibility and ablation (e.g., pretraining corpora, safety filters, refusal policies); need open baselines to enable causal analyses.
- Limited analysis of instruction semantics: The “Unknown” option boosts accuracy but semantics of abstention are not explored; requires experiments with calibrated “insufficient information” options and decision-aware scoring.
- Limited diversity and fairness analysis: No subgroup performance by demographics, imaging devices, institutions, or languages; fairness, domain shift, and cross-site generalization remain untested.
- Image quality and acquisition variability: Robustness to real-world imaging artifacts (noise, compression, cropping, orientation, scanner differences) is not stress-tested.
- Multi-turn clinical reasoning: Single-turn MCQ framing omits clarification queries, hypothesis updating, and tool use; effects of interactive workflows are unknown.
- Long-form generation robustness: Report-generation metrics are listed but not stress-tested for hallucinations, omission errors, and factual grounding under perturbations.
- Comparison to specialist models: Limited examination of domain-specific systems and fine-tuned medical LMMs; unclear whether specialization mitigates shortcut behaviors.
- Scaling effects and mechanisms: No analysis of how model size, training data composition, or architecture choices affect shortcut reliance and robustness trends.
- Causal diagnostics of shortcuts: Lacks mechanistic or interpretability studies (e.g., representation probing, counterfactual tracing) to identify when and why models ignore images or fabricate reasoning.
- Regulatory alignment: No mapping from robustness findings to actionable regulatory criteria or release standards (e.g., minimum acceptable abstention/robustness thresholds).
- Public release of annotations and code: Unclear availability of item-level labels, perturbation scripts, and prompts needed for replication and extension by the community.
Practical Applications
Immediate Applications
The following items can be deployed now using the paper’s stress-testing methods, clinician-guided rubric, and reporting practices, to improve evaluation, procurement, deployment, and education around multimodal medical AI.
- Robustness stress-testing as a pre-deployment gate for medical AI
- Description: Integrate Tests T1–T6 (modality sensitivity/necessity, format/distractor perturbations, visual substitution, reasoning audit) into model evaluation before clinical pilots and updates.
- Sectors: Healthcare delivery, health IT, software/ML tooling, medical imaging.
- Tools/products/workflows: “MedStressLab” evaluation harness (internal or open-source) that automates T1–T6; CI/CD hooks in MLflow/W&B to fail builds if mean robustness score drops; vendor scorecards that report accuracy and mean robustness score side-by-side.
- Assumptions/dependencies: Access to representative benchmark subsets; governance approval to add gates in MLOps; compute for systematic perturbations; versioned prompts and datasets to ensure reproducibility.
- Clinician-guided benchmark profiling for fit-for-purpose evaluations
- Description: Use the paper’s 10-dimension rubric to tag existing test sets along reasoning and visual complexity, then match evaluation to the intended clinical use (e.g., diagnosis vs. description; visually essential vs. text-solvable).
- Sectors: Healthcare, academia, benchmark maintainers.
- Tools/products/workflows: “MedBench-Profiler” script to score datasets and generate profile plots; updated model cards with per-axis performance rather than single leaderboard numbers.
- Assumptions/dependencies: Clinician time (triad reviews) for rubric scoring; inter-rater agreement tracking; dataset licenses.
- Procurement and contracting language that requires robustness evidence
- Description: Health systems and payers require vendors to report modality sensitivity/necessity results, perturbation robustness, and refusal behavior under missing modalities, not just accuracy.
- Sectors: Healthcare administration, payers, legal/policy, health IT procurement.
- Tools/products/workflows: RFP templates mandating mean robustness score thresholds; disaggregated results by benchmark profile; attestations on data leakage checks and contamination controls.
- Assumptions/dependencies: Organizational policy updates; vendor cooperation; clear thresholds and acceptance criteria.
- Safer product UX: calibrated abstention and uncertainty when critical input is missing
- Description: Configure models to abstain (or require secondary review) on visually-required tasks when images are absent or low quality; treat “Unknown” as a legitimate option with calibrated selective prediction, not as a weakened distractor.
- Sectors: Healthcare software, consumer health apps, imaging platforms.
- Tools/products/workflows: Uncertainty prompts; abstain buttons; quality checks on modality presence; selective prediction thresholds defined by T1–T2; audit logs when abstaining overrides are made.
- Assumptions/dependencies: Policy approval for abstention behavior; UX changes; monitoring to ensure abstentions do not degrade workflow throughput.
- Dataset hygiene and artifact reduction
- Description: Use T3–T5 to detect and mitigate shortcut cues (answer-position bias, frequent distractors, visual-answer pair leakage); rebalance option orders and diversify distractors.
- Sectors: Academia, benchmark maintainers, model developers.
- Tools/products/workflows: Automated detection of position bias; distractor generators; counterfactual visual substitution checks before dataset release.
- Assumptions/dependencies: Access to item authoring pipelines; acceptance of slightly lower raw accuracy as shortcuts are removed.
- Publication and peer-review checklists
- Description: Journals and conferences request authors to report stress-test outcomes, reasoning audits, and benchmark profiles alongside accuracy.
- Sectors: Academia, scientific publishing.
- Tools/products/workflows: Standardized appendix items for T1–T6; reviewer rubrics to flag modality negligence or hallucinated reasoning.
- Assumptions/dependencies: Editorial policy updates; community norms.
- Post-market surveillance using canary stress sets
- Description: Maintain a small, versioned “canary” panel of stress items to monitor drift and brittleness in production models.
- Sectors: Healthcare systems, vendors, MLOps.
- Tools/products/workflows: Scheduled evaluation jobs; dashboards tracking accuracy vs. robustness drift; rollback triggers.
- Assumptions/dependencies: Data governance for safe test injection; monitoring infrastructure.
- Education and training with failure exemplars
- Description: Teach clinicians and trainees common LMM failure modes (fabricated rationales, misgrounded perception, format sensitivity) using real examples from T6 audits.
- Sectors: Medical education, continuing education.
- Tools/products/workflows: Case libraries; simulation sessions where trainees adjudicate AI rationales; competency checklists.
- Assumptions/dependencies: De-identified examples; faculty time.
- Cross-domain adoption of the perturbation suite
- Description: Apply T1–T5 analogs to other multimodal domains (e.g., pathology, ophthalmology, industrial inspection).
- Sectors: Robotics/inspection, manufacturing, energy, autonomous systems.
- Tools/products/workflows: Sensor-modality necessity tests; distractor-like alternative hypotheses; counterfactual sensor substitutions.
- Assumptions/dependencies: Domain experts to define “visually required” items; sector-specific safety policies.
- Consumer guidance and labeling
- Description: Communicate that high benchmark scores do not equal clinical reliability; encourage apps to display model limitations and when abstention is triggered.
- Sectors: Daily life, consumer health.
- Tools/products/workflows: In-app labels; FAQs explaining when images are essential and why the model might refuse.
- Assumptions/dependencies: Regulatory alignment on labeling; UX space for disclosures.
Long-Term Applications
The following items require further research, scaling, standards-setting, or regulatory action before broad deployment.
- Medical Robustness Index and certification
- Description: Establish a standardized, third-party audited index (built on mean robustness score and rubric axes) that certifies models for specific clinical claims.
- Sectors: Healthcare, standards bodies, regulators.
- Tools/products/workflows: Evaluation-as-a-service labs; accreditation processes; public registry of certified claims and scope.
- Assumptions/dependencies: Consensus on metrics and thresholds; test-set governance; legal frameworks for certification.
- Dynamic, contamination-resistant benchmarks
- Description: Rotating items, secure enclaves, and synthetic counterfactuals to reduce memorization and co-occurrence shortcuts; support longitudinal and multi-view tasks.
- Sectors: Academia, benchmark maintainers, regulators.
- Tools/products/workflows: Secure evaluation servers; item pools with controlled exposure; synthetic visual substitution generators; leakage-detection reports.
- Assumptions/dependencies: Funding for ongoing curation; compute for generation; privacy-preserving infrastructure.
- Training objectives for grounded, faithful reasoning
- Description: Losses and supervision that tie rationales to verifiable visual evidence; penalties for hallucinated perception; reinforcement from human feedback that rewards abstention under missing modalities.
- Sectors: AI research, model vendors.
- Tools/products/workflows: Grounding verification modules; counterfactual-consistency training (robust to T5 substitutions); selective prediction and calibration (ECE, risk-based thresholds).
- Assumptions/dependencies: High-quality rationale annotations; scalable grounding checks; alignment between training incentives and safety goals.
- Regulatory guidance for modality-dependent claims and abstention
- Description: Require explicit disclosure of tasks that are text-solvable vs. visually essential; mandate selective abstention policies and user-visible uncertainty.
- Sectors: Policy/regulation, healthcare delivery.
- Tools/products/workflows: Labeling standards; post-market performance monitoring (including T1–T6 reports); enforcement mechanisms.
- Assumptions/dependencies: Harmonization across agencies; clear liability frameworks for abstentions and overrides.
- Post-deployment stress-testing and drift governance
- Description: Continuous evaluation under real-world shifts (image quality, device differences, population changes) with automated alerts when robustness degrades.
- Sectors: Health systems, vendors, MLOps.
- Tools/products/workflows: Live data shadow tests; domain shift detectors; remediation playbooks (recalibration, retraining, guardrail tightening).
- Assumptions/dependencies: Data-sharing agreements; safe evaluation pipelines; monitoring budgets.
- Human-AI collaboration protocols grounded in brittleness signals
- Description: Route cases to human experts when T1–T5 indicators suggest shortcut-prone scenarios; dynamically increase human oversight for visually essential or high-uncertainty items.
- Sectors: Healthcare operations, radiology/pathology services.
- Tools/products/workflows: Case-triage policies; escalation logic in PACS/EHR; attribution of final responsibility and documentation of AI rationale usage.
- Assumptions/dependencies: Workflow redesign; acceptance by clinicians; auditability.
- Sector-general multimodal safety standards
- Description: Export the framework to other industries where decisions depend on fusing text, images, and structured signals (e.g., finance KYC with document images, insurance claims, autonomous drones).
- Sectors: Finance, insurance, robotics, public safety.
- Tools/products/workflows: Domain-specific modality necessity tests; adversarial counterfactuals; refusal/abstention policies.
- Assumptions/dependencies: Domain expertise; adaptation of rubrics; sector regulators’ buy-in.
- Data standards and metadata for benchmark profiling
- Description: Embed rubric-aligned metadata in dataset standards (e.g., DICOM, FHIR/HL7 extensions) to indicate reasoning and visual demands.
- Sectors: Standards bodies, EHR vendors, imaging system vendors.
- Tools/products/workflows: Schema extensions; dataset publishing guidelines; validators that check completeness of benchmark profiles.
- Assumptions/dependencies: Backward compatibility; vendor adoption cycles.
- Prospective clinical trials incorporating stress tests
- Description: Design pragmatic trials that measure clinical utility and safety while also tracking T1–T6 performance over time and across sites.
- Sectors: Clinical research, regulators, payers.
- Tools/products/workflows: Trial protocols with co-primary endpoints (utility + robustness); site heterogeneity analyses; reimbursement policy tied to robustness maintenance.
- Assumptions/dependencies: Funding; IRB approvals; multi-site coordination.
- Counterfactual data generation at scale
- Description: Systematic creation of visual substitutions aligned to distractor hypotheses to pressure-test visual-textual integration during both training and evaluation.
- Sectors: AI research, data vendors, imaging manufacturers.
- Tools/products/workflows: Generative pipelines (e.g., synthetic radiology images) with clinical validation; automated counterfactual matching to answer options.
- Assumptions/dependencies: Validated generative models; clinical reviewer capacity; controls to avoid synthetic artifacts becoming new shortcuts.
Notes on feasibility across applications
- Many benefits depend on community norms shifting from single-number leaderboards to multi-axis, stress-tested reporting.
- Vendor transparency and data provenance checks are crucial to detect pretraining contamination and avoid overestimating capability.
- Some models’ refusal behavior under missing modalities may lower benchmark accuracy but improve real-world safety; evaluation policies must not penalize safe abstentions.
- Compute costs and clinician time are non-trivial; shared toolkits and standardized test panels can reduce burden.
Glossary
- Adversarial input: Inputs intentionally or unintentionally degraded or misleading that test model robustness. "degraded, incomplete, or adversarial input"
- Angioid streaks: Breaks in Bruch’s membrane seen on retinal imaging, associated with systemic conditions. "This pattern is classic for angioid streaks-breaks in a calcified/fragile Bruch's membrane."
- Board-certified clinicians: Clinicians who have met specialty certification standards, used here as expert annotators. "Each benchmark was independently annotated by three board-certified clinicians per axis"
- Bruch's membrane: A thin layer of the eye’s retina whose disruption produces angioid streaks. "classic for angioid streaks-breaks in a calcified/fragile Bruch's membrane."
- Chain-of-Thought (CoT): A prompting technique that elicits step-by-step reasoning in model outputs. "Chain-of- Thought (CoT) [11] prompting"
- Clinical vignette: A short narrative of patient context often paired with images in medical questions. "pairing a brief clinical vignette"
- Co-occurrence patterns: Statistical associations between features (e.g., symptoms and diagnoses) that models may exploit. "co-occurrence patterns, or memorized question-answer pairs"
- Counterfactual: A minimally altered scenario used to test whether models revise predictions appropriately. "This creates a new 'counterfactual' version"
- Differential diagnosis: The process of distinguishing among multiple possible conditions explaining a presentation. "Differential diagnosis or staged inference"
- Distractor: An incorrect answer option designed to challenge the model’s ability to choose the correct answer. "We progressively replace 1-4 distractor options with irrelevant choices"
- Distributional shift: Changes in input data distribution that can degrade model performance. "brittleness under distributional shift, missing data, or subtle perturbations."
- Elimination-based reasoning: Strategy of ruling out options using superficial cues rather than deep understanding. "highlighting their elimination-based reasoning strategy."
- Fabricated reasoning: Confident but incorrect or invented rationales that do not reflect true evidence. "fabricate convincing yet flawed reasoning."
- Fleiss' K: A statistical measure of inter-annotator agreement for categorical ratings among multiple raters. "measure inter-annotator agreement using Fleiss' K [19]."
- Fragility value: A normalized measure of instability for a model under a specific stress test. "we compute a fragility value fi(m) € [0,1]."
- Frontier models: The most advanced large-scale AI systems at the cutting edge of capability. "Large frontier models like GPT-5 now achieve top scores on medical bench- marks."
- Hallucinated Perception: A failure mode where models describe visual features that are not present. "Correct Answer but Hallucinated Perception"
- Heliotrope rash: Violaceous discoloration of the eyelids characteristic of dermatomyositis. "classic 'heliotrope rash.'"
- Inter-annotator agreement: The degree to which different experts provide consistent labels or ratings. "measure inter-annotator agreement using Fleiss' K [19]."
- JAMA: A benchmark derived from the Journal of the American Medical Association used for multimodal QA. "On JAMA, absolute scores were higher and changes smaller:"
- LMMs (Large Multimodal Models): AI systems that process and integrate text and images (and other modalities). "Stress tests reveal hidden fragilities in LMMs on multimodal medical tasks."
- Memorized question-answer pairs: Learned associations from training data that can be recalled without understanding. "memorized question-answer pairs"
- MIMIC-CXR: A large chest X-ray dataset used for report generation and imaging tasks. "performance evaluation on chest X-ray report generation using the MIMIC-CXR dataset."
- Modality necessity: The extent to which a specific modality (e.g., image) is required to answer a question correctly. "Stress Test 2: Modality necessity"
- Modality sensitivity: How much performance changes when a modality (e.g., image) is removed or degraded. "T1: Modality Sensitivity"
- Multimodal medical benchmarks: Evaluation datasets combining text and medical images to test model capability. "On multimodal medical benchmarks [3, 4], leading models retained most of their original accuracy even when images were removed."
- Multimodal sensitivity: Sensitivity of performance to the presence or absence of multiple input modalities. "Stress Test 1: Multimodal sensitivity"
- OmniMedVQA: A medical visual question answering benchmark with relatively low reasoning and visual complexity. "increasing reasoning strength on OmniMedVQA offers only minor improvements"
- Ordinal scale: A ranked scoring scheme (e.g., 3-point) used for rubric-based annotation. "using a 3-point ordinal scale."
- Osler nodes: Painful, raised lesions on fingers or toes indicative of endocarditis. "Osler nodes, Janeway lesions"
- Path-VQA: A visual question answering benchmark focused on pathology images. "Path-VQA"
- PMC-VQA: A biomedical VQA dataset constructed from PubMed Central articles. "PMC-VQA [17]"
- Priors (frequency priors): Assumptions about the likelihood of outcomes learned from data frequencies. "frequency priors, co-occurrence patterns, or memorized question-answer pairs"
- RadCliQ-v1: A metric/framework for evaluating radiology report quality. "RadCliQ-v1"
- RadGraph F1: A structured information extraction metric for radiology reports assessing graph-level correctness. "RadGraph F1"
- Reinforcement learning approaches: Training methods that optimize behavior via rewards, potentially at token-level. "Reinforcement learning approaches may optimize for token- level reward signals rather than faithful reasoning."
- ReXrank leaderboard: A standardized platform for benchmarking radiology report generation models. "based on the ReXrank leaderboard [20]"
- Robustness score: A normalized measure (0–1) aggregating stability across stress tests. "The mean robustness score was obtained by averaging across the five tests"
- Shortcut behavior: Model tendencies to exploit superficial patterns to get correct answers without true understanding. "shortcut behavior"
- Shortcut learning: Learning spurious heuristics that correlate with correct answers instead of genuine comprehension. "This is shortcut learning, not medical understanding."
- Spurious cues: Irrelevant signals or artifacts that models use to predict answers. "strip away spurious cues"
- Stress tests: Targeted evaluations that perturb inputs to reveal brittleness and shortcut reliance. "Our findings call for a fundamental reevaluation... we apply a series of targeted stress tests"
- Visual grounding: Tying explanations or decisions to actual visual evidence in the image. "examining their factuality, visual grounding, and alignment with final answers."
- Visual salience: The prominence or conspicuousness of visual features affecting diagnostic difficulty. "the new images vary in visual salience or diagnostic ambiguity."
- Visual substitution: Replacing an image with one supporting a different option to test integration of vision and text. "Stress Test 5: Visual Substitu- tion"
- Visual-textual integration: The ability to jointly reason over images and text to reach consistent conclusions. "rather than robust visual-textual integration"
- Visual-answer pairings: Learned associations linking particular visuals to specific answer labels. "rely on learned visual-answer pair- ings rather than interpreting visual evidence in context."
- VQA-RAD: A medical visual question answering dataset focused on radiology images. "VQA-RAD [12]"
Collections
Sign up for free to add this paper to one or more collections.