Leveraging Prompt-Learning for Structured Information Extraction from Crohn's Disease Radiology Reports in a Low-Resource Language (2405.01682v2)
Abstract: Automatic conversion of free-text radiology reports into structured data using NLP techniques is crucial for analyzing diseases on a large scale. While effective for tasks in widely spoken languages like English, generative LLMs typically underperform with less common languages and can pose potential risks to patient privacy. Fine-tuning local NLP models is hindered by the skewed nature of real-world medical datasets, where rare findings represent a significant data imbalance. We introduce SMP-BERT, a novel prompt learning method that leverages the structured nature of reports to overcome these challenges. In our studies involving a substantial collection of Crohn's disease radiology reports in Hebrew (over 8,000 patients and 10,000 reports), SMP-BERT greatly surpassed traditional fine-tuning methods in performance, notably in detecting infrequent conditions (AUC: 0.99 vs 0.94, F1: 0.84 vs 0.34). SMP-BERT empowers more accurate AI diagnostics available for low-resource languages.
- Leveraging gpt-4 for post hoc transformation of free-text radiology reports into structured reporting: A multilingual feasibility study. Radiology, 307(4).
- Language models are few-shot learners. In Advances in Neural Information Processing Systems, volume 33, pages 1877–1901. Curran Associates, Inc.
- Consensus recommendations for evaluation, interpretation, and utilization of computed tomography and magnetic resonance enterography in patients with small bowel crohn’s disease. Radiology, 286(3):776–799.
- Deep learning–based assessment of oncologic outcomes from natural language processing of structured radiology reports. Radiology: Artificial Intelligence, 4(5).
- Ai-luminating artificial intelligence in inflammatory bowel diseases: A narrative review on the role of ai in endoscopy, histology, and imaging for ibd. Inflammatory Bowel Diseases.
- Clinicalbert: Modeling clinical notes and predicting hospital readmission. arXiv:1904.05342.
- Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Computing Surveys, 55(9):1–35.
- Exploring the boundaries of gpt-4 in radiology. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics.
- Ct and mri evaluations in crohn’s complications: A guide for the radiologist. Academic Radiology, 29(8):1206–1227.
- Practical guide to natural language processing for radiology. RadioGraphics, 41(5):1446–1453.
- Feasibility of using the privacy-preserving large language model vicuna for labeling radiology reports. Radiology, 309(1).
- Language models are unsupervised multitask learners. OpenAI blog, 1(8):9.
- Timo Schick and Hinrich Schütze. 2021a. Exploiting cloze-questions for few-shot text classification and natural language inference. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume. Association for Computational Linguistics.
- Timo Schick and Hinrich Schütze. 2021b. It’s not just size that matters: Small language models are also few-shot learners. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Association for Computational Linguistics.
- On the Stratification of Multi-label Data, page 145–158. Springer Berlin Heidelberg.
- Vitaly Shalumov and Harel Haskey. 2023. Hero: Roberta and longformer hebrew language models. arXiv preprint arXiv:2304.11077.
- Combining automatic labelers and expert annotations for accurate radiology report labeling using bert. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). Association for Computational Linguistics.
- NSP-BERT: A prompt-based few-shot learner through an original pre-training task —— next sentence prediction. In Proceedings of the 29th International Conference on Computational Linguistics, pages 3233–3250, Gyeongju, Republic of Korea. International Committee on Computational Linguistics.
- Performance of multiple pretrained bert models to automate and accelerate data annotation for large datasets. Radiology: Artificial Intelligence, 4(4).
- Crohn’s disease. The Lancet, 389(10080):1741–1755.
- Entailment as few-shot learner.
- Radbert: Adapting transformer-based language models to radiology. Radiology: Artificial Intelligence, 4(4).
- Liam Hazan (5 papers)
- Gili Focht (2 papers)
- Naama Gavrielov (2 papers)
- Roi Reichart (82 papers)
- Talar Hagopian (2 papers)
- Mary-Louise C. Greer (1 paper)
- Ruth Cytter Kuint (1 paper)
- Dan Turner (3 papers)
- Moti Freiman (33 papers)