Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Improving Zero-Shot Detection of Low Prevalence Chest Pathologies using Domain Pre-trained Language Models (2306.08000v1)

Published 13 Jun 2023 in physics.med-ph, cs.CL, cs.CV, cs.LG, and eess.IV

Abstract: Recent advances in zero-shot learning have enabled the use of paired image-text data to replace structured labels, replacing the need for expert annotated datasets. Models such as CLIP-based CheXzero utilize these advancements in the domain of chest X-ray interpretation. We hypothesize that domain pre-trained models such as CXR-BERT, BlueBERT, and ClinicalBERT offer the potential to improve the performance of CLIP-like models with specific domain knowledge by replacing BERT weights at the cost of breaking the original model's alignment. We evaluate the performance of zero-shot classification models with domain-specific pre-training for detecting low-prevalence pathologies. Even though replacing the weights of the original CLIP-BERT degrades model performance on commonly found pathologies, we show that pre-trained text towers perform exceptionally better on low-prevalence diseases. This motivates future ensemble models with a combination of differently trained LLMs for maximal performance.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (8)
  1. Publicly available clinical BERT embeddings. CoRR, abs/1904.03323, 2019. URL http://arxiv.org/abs/1904.03323.
  2. Making the most of text semantics to improve biomedical vision–language processing, 2022. URL https://arxiv.org/abs/2204.09817.
  3. A survey on deep learning in medical image analysis. Medical Image Analysis, 42:60–88, 2017. ISSN 1361-8415. https://doi.org/10.1016/j.media.2017.07.005. URL https://www.sciencedirect.com/science/article/pii/S1361841517301135.
  4. Transfer learning in biomedical natural language processing: An evaluation of BERT and elmo on ten benchmarking datasets. CoRR, abs/1906.05474, 2019. URL http://arxiv.org/abs/1906.05474.
  5. Learning transferable visual models from natural language supervision. CoRR, abs/2103.00020, 2021. URL https://arxiv.org/abs/2103.00020.
  6. Chexnet: Radiologist-level pneumonia detection on chest x-rays with deep learning, 2017. URL https://arxiv.org/abs/1711.05225.
  7. Chexbert: Combining automatic labelers and expert annotations for accurate radiology report labeling using BERT. CoRR, abs/2004.09167, 2020. URL https://arxiv.org/abs/2004.09167.
  8. Expert-level detection of pathologies from unannotated chest x-ray images via self-supervised learning. Nature Biomedical Engineering, pages 1–8, 09 2022. 10.1038/s41551-022-00936-9.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Aakash Mishra (3 papers)
  2. Rajat Mittal (42 papers)
  3. Christy Jestin (1 paper)
  4. Kostas Tingos (1 paper)
  5. Pranav Rajpurkar (69 papers)
Citations (7)