Medical QA: Challenges & Advances
- Medical domain question answering is the task of generating accurate, context-aware responses to clinical queries using text, visual, and spoken formats.
- It employs multi-stage, retrieval-augmented architectures and domain-specific language models to tackle complex reasoning and multi-modal challenges.
- Recent advances have enhanced dataset quality, evaluation metrics, and knowledge injection methods to reduce epistemic risks and improve reliability.
Medical domain question answering (QA) is the task of producing accurate, concise, and context-aware answers to natural-language questions posed by clinicians, patients, researchers, or students, across a spectrum of formats (multiple choice, extractive, open-ended, visual, and spoken). Medical QA systems must address substantial challenges in domain adaptation, coverage of complex medical reasoning, multi-modal and multilingual input, and the management of high epistemic risk. Recent advances have led to the development of large-scale datasets, robust retrieval-augmented architectures, and sophisticated evaluation protocols to benchmark and improve model performance across diverse languages and clinical subfields.
1. Problem Formulations, Task Types, and Evaluation
Medical QA encompasses a range of sub-tasks, including multiple-choice question answering (MCQA), extractive (span-based) question answering (EQA), open-ended generation, visual QA (Med-VQA), and spoken QA (SQA). Each sub-task targets different use-cases—from standardized exams (MCQA) to clinical information retrieval and patient support.
Multiple-Choice QA (MCQA):
- Systems predict the correct answer(s) from a fixed set of candidates. Datasets such as MedMCQA (Pal et al., 2022), FrenchMedMCQA (Labrak et al., 2023), MediQAl (Bazoge, 28 Jul 2025), and MedQA (Jin et al., 2020) provide large-scale, domain-authentic benchmarks with fine-grained category and difficulty labeling.
- Evaluation metrics: Accuracy (single-answer MCQ), Hamming score and Exact Match Ratio (EMR) for multi-label MCQA, as well as per-question precision, recall, and F for multi-answer tasks (Labrak et al., 2023).
Extractive QA (Medical-EQA):
- The system extracts a span from an often lengthy passage as the answer to a question (e.g., "What is the recommended dose of remdesivir...?"). Metrics include Exact Match (EM) and token-overlap F (Sengupta et al., 2023).
Open-Ended and Generative QA:
- Answers are generated as free text, often evaluated with BERTScore, ROUGE, BLEU, METEOR, or LLM-as-Judge metrics (Romano et al., 21 Oct 2025, Bazoge, 28 Jul 2025).
Visual and Spoken QA:
- Visual QA involves answering based on images (e.g., radiographs), requiring vision-LLMs; evaluation typically uses accuracy on both closed and open question formats (Ha et al., 2024, Canepa et al., 2023).
- Spoken QA operates on audio-form questions, preferably with end-to-end models that map directly from audio to answer candidates without transcription, evaluated by accuracy (Labrak et al., 2024).
2. Data Resources and Dataset Construction
The development of robust medical QA relies fundamentally on diverse and high-quality datasets. Major corpora include:
Written QA Datasets:
| Dataset | Language(s) | Size | Task Types | Source Domain |
|---|---|---|---|---|
| MedMCQA (Pal et al., 2022) | English | ≈194k MCQ | MCQA, Explanation | Entrance exams |
| FrenchMedMCQA (Labrak et al., 2023) | French | 3,105 MCQA | Single/Multiple MCQA | Pharmacy exams |
| MediQAl (Bazoge, 28 Jul 2025) | French | 32,603 MCQ/OEQ | MCQU, MCQM, OEQ | Medical exams |
| MedQA (Jin et al., 2020) | En, Zh-cn, Zh-tw | 61k MCQ | Multi-lingual MCQA | Board exams |
| RoMedQA (Rogoz et al., 22 Aug 2025) | Romanian | 102,646 QA | Extractive, Reasoning | Oncology EHRs |
| IMB (Romano et al., 21 Oct 2025) | Italian | 782k QA, 26k MCQ | Open/Multi-choice | Forums, Exams |
| ECN-QA (Khlaut et al., 2024) | French/En | 5,531 QA | MCQ, Progressive | Clinical vignettes |
Visual and Spoken QA Datasets:
- SLAKE 1.0, VQA-RAD, and VQA-Med for visual QA (Ha et al., 2024, Canepa et al., 2023).
- SQA benchmarks derived from MCQA datasets, synthesized into audio (Labrak et al., 2024).
Datasets are constructed with rigorous data cleaning, anonymization, language adaptation, and domain expert validation, often spanning multiple clinical specialties and reasoning types. Newer strategies include the generation of synthetic clinical scenarios via LLM prompting, curriculum design for exam-style progression (progressive questions), and the annotation of cognitive demands (e.g., “reasoning” vs. “factual recall”) (Bazoge, 28 Jul 2025, Khlaut et al., 2024).
3. Model Architectures: Retrieval, Reading, and Generation
State-of-the-art medical QA systems are predominantly based on multi-stage, retrieval-augmented architectures, combined with Transformer-based deep neural networks for answer prediction or generation.
Retriever-Reader Paradigm:
- An initial retriever module (BM25, Dense Passage Retrieval, DPR) identifies relevant passages from large biomedical corpora (PubMed, HAL, clinical notes) (Gupta, 2023, Hassan et al., 5 Dec 2025, Baksi, 2021).
- The reader (extractive BERT-style model for span selection, or generative T5/BART-style for free-form output) then processes the concatenated question, answer candidates, and retrieved context (Pal et al., 2022, Labrak et al., 2023).
- Incorporating multiple retrieved passages, segment-level and semantic filtering, and explicit citation grounding further boosts performance and reliability (Hassan et al., 5 Dec 2025, Levy et al., 2021).
Domain Adaptation and Fine-Tuning:
- Best performance is consistently achieved by domain-specialized encoders (BioBERT, PubMedBERT, RadBloomz, domain-tuned CLIP) and LLMs further pre-trained or fine-tuned on domain/corpus-specific tasks (Labrak et al., 2023, Pal et al., 2022, Ha et al., 2024).
- Parameter-efficient adaptation methods such as LoRA are widely used for fine-tuning LLMs (e.g., LLaMA-2, Falcon) in the biomedical specificity regime (Hassan et al., 5 Dec 2025).
- For low-resource languages and domains, supervised fine-tuning and domain/language-specific adaptation reliably outperform zero-shot transfer even for large LLMs (Rogoz et al., 22 Aug 2025, Romano et al., 21 Oct 2025).
Generative and Multi-Agent Models:
- Generative approaches, especially when fine-tuned in a sequence-to-sequence format, handle multi-answer and open-ended outputs natively (e.g., BART-base, GPT-4, Llama3.1-70B) (Labrak et al., 2023, Yang et al., 2024).
- Hybrid multi-agent pipelines, where different LLM instances specialize in query analysis, case generation, and expert report synthesis, further boost accuracy and interpretability in multi-choice settings (Yang et al., 2024).
Knowledge Injection:
- Resource-efficient techniques for injecting structured medical knowledge into otherwise open-domain LMs leverage knowledge graph embeddings (e.g., UMLS/Metathesaurus) projected into LM embedding spaces, sometimes matching domain-pretrained models in performance (Sengupta et al., 2024).
- Recognizing Question Entailment (RQE) serves as a pipeline to find semantically related previously answered questions for improved coverage, especially when combined with trusted curated QA banks (Abacha et al., 2019).
4. Multimodal, Multilingual, and Spoken Medical QA
Beyond text-only systems, medical QA now extends to multimodal and cross-language environments:
Visual QA:
- Fusion models integrate frozen, domain-adapted vision encoders (BiomedCLIP ViT) and radiology-pretrained language decoders (RadBloomz-7b), with parameter-efficient adapters (LoRA, Query Transformer) incrementally trained across image captioning and biomedical VQA tasks (Ha et al., 2024, Canepa et al., 2023).
- Fine-grained accuracy is reported both for closed (yes/no/multiclass) and open questions, with error analyses emphasizing synonymy, paraphrase, and spatial reasoning difficulty.
Spoken QA:
- Zero-shot, end-to-end (E2E) systems process audio with a joint audio–text encoder (CLAP, Whisper), eliminating the transcription bottleneck inherent to cascaded ASR-to-LLM pipelines and achieving comparable or slightly better accuracy with an order-of-magnitude less compute (Labrak et al., 2024).
- Evaluation is performed over synthesized spoken versions of major QA sets, highlighting the need for domain-adapted audio models and future work on in-context SQA exemplars and real-world speech variability.
Multilinguality:
- Robust medical QA systems (FrenchMedMCQA (Labrak et al., 2023), MediQAl (Bazoge, 28 Jul 2025), RoMedQA (Rogoz et al., 22 Aug 2025), IMB (Romano et al., 21 Oct 2025), MedQA (Jin et al., 2020)) confirm that domain-tuned models (even in English) often outperform monolingual LMs in low-resource languages, though language-specific adaptation remains essential for clinical reliability.
- LLMs in zero-shot fail to generalize on out-of-language QA, with fine-tuning and domain lexicon adaptation necessary to close the linguistic gap (Rogoz et al., 22 Aug 2025, Romano et al., 21 Oct 2025).
5. Evaluation Methodologies, Error Modes, and Practical Recommendations
Medical QA evaluation relies on a mixture of metrics attuned to each task:
- MCQA: Accuracy, Hamming score, Exact Match Ratio (EMR), precision/recall/F for multi-label predictions (Labrak et al., 2023, Bazoge, 28 Jul 2025).
- Extractive QA: Exact Match (EM), token-level F, per-span scoring for multiple non-overlapping answer spans (Sengupta et al., 2023, Sengupta et al., 2024).
- Open-ended QA: Text similarity (BERTScore, ROUGE, BLEU, METEOR), sometimes scored by LLMs for human-comparable assessment (Romano et al., 21 Oct 2025, Bazoge, 28 Jul 2025).
- Visual/Spoken QA: Answer classification accuracy, fuzzy-matching for open and synonymic answers (Ha et al., 2024, Labrak et al., 2024).
Error Analysis:
- Models frequently miss multi-hop reasoning chains, make errors on arithmetic or dose-calculation questions, and oversimplify in multi-label/multi-step inferences (Pal et al., 2022, Bazoge, 28 Jul 2025).
- Hallucinations—answers not grounded in the retrieved context—are a common failure mode for generative LLMs; retrieval grounding and secondary verification can reduce hallucination by up to 60% (Hassan et al., 5 Dec 2025).
- Negation handling, distractor confusion, and reasoning about rare or underrepresented subjects are other observed challenges (Labrak et al., 2023, Bazoge, 28 Jul 2025).
- Visual QA suffers from limited annotated data and overfitting, favoring shallower, parameter-light models and non-autoregressive output heads (Canepa et al., 2023).
Practical Recommendations:
- Use domain-adapted or in-domain pretrained models whenever possible; fine-tuning generic LLMs is insufficient.
- Dense retrieval on domain-specific corpora (PubMed, HAL, specialized forums/guidelines) yields higher-quality context than general sources.
- For low-resource language or specialty domains, invest in expert-validated, anonymized, and paraphrased clinical QA datasets; split by patient/case to avoid leakage (Rogoz et al., 22 Aug 2025, Romano et al., 21 Oct 2025).
- Employ parameter-efficient fine-tuning (LoRA), moderate context windows, and prompt engineering for optimal performance on both extractive and generative systems (Hassan et al., 5 Dec 2025, Rogoz et al., 22 Aug 2025).
- Integrate explicit citation or retrieval-based grounding to mitigate hallucination, with audit trails for clinical deployment (Hassan et al., 5 Dec 2025).
- Introduce multi-modal fusion or hybrid extractive–generative pipelines where appropriate; consider RAG for “open-book” QA settings (Ha et al., 2024, Khlaut et al., 2024).
6. Recent Advances and Future Directions
Notable directions at the research frontier include:
- Target-Oriented Pretraining (TOP-Training): Pretraining using synthetic corpora that match the entity and style distribution of target EQA datasets delivers higher downstream extractive accuracy with minimal text generation cost; care is needed to validate LLM-generated documents (Sengupta et al., 2023).
- Knowledge-Augmented Data Generation: CVAE-based and LLM-augmented question drafting pipelines expand small/specialist datasets and boost accuracy, especially in low-resource or underrepresented topics (Shen et al., 2018, Khlaut et al., 2024).
- Hybrid Retrieval + Entailment Models: Recognizing Question Entailment (RQE) reliably boosts domain QA performance, with lightweight classifiers outperforming deep models on clinical entailment, particularly when paired with trusted-source QA banks and hybrid IR/reranker pipelines (Abacha et al., 2019).
- Multimodal and Deep Reasoning: Parameter-efficient adaptation enables fusion models for visual QA and EHRs, but limited label availability remains a bottleneck (Ha et al., 2024).
- Resource-Efficient Knowledge Injection: Mapping structured knowledge graph embeddings to LM spaces compensates for lack of biomedical pretraining in general-purpose LMs with minimal compute (Sengupta et al., 2024).
- Spoken QA and Low-Resource Domains: E2E spoken QA and robust low-resource language adaptations will rely on synthetic task construction, retrieval grounding, and active human-in-the-loop validation (Labrak et al., 2024, Rogoz et al., 22 Aug 2025, Romano et al., 21 Oct 2025).
A clear trend is the convergence of modular retrieval-grounded pipelines, efficient domain adaptation (via synthetic data, knowledge graphs, or targeted pretraining), rigorous multi-dimensional evaluation, and the deliberate curation of multilingual, multi-specialty, and multi-format datasets. Open challenges remain in multi-modal clinical reasoning, robust generalization to new specialties and languages, context-sensitive error detection, trustworthy reasoning, and comprehensive explainability.
Key References:
- FrenchMedMCQA (Labrak et al., 2023)
- MedMCQA (Pal et al., 2022)
- PubMedQA/MedQA (Jin et al., 2020)
- MediQAl (Bazoge, 28 Jul 2025)
- IMB (Romano et al., 21 Oct 2025)
- RoMedQA (Rogoz et al., 22 Aug 2025)
- LLM-MedQA (Yang et al., 2024)
- RAG architectures (Hassan et al., 5 Dec 2025)
- Knowledge graph hybridization (Sengupta et al., 2024)
- Synthetic data/augmentation (Shen et al., 2018, Khlaut et al., 2024)
- Visual/spoken QA (Ha et al., 2024, Canepa et al., 2023, Labrak et al., 2024)
- TOP-Training (Sengupta et al., 2023)
- Recognizing Question Entailment (Abacha et al., 2019)