RadLing: Towards Efficient Radiology Report Understanding (2306.02492v1)
Abstract: Most natural language tasks in the radiology domain use LLMs pre-trained on biomedical corpus. There are few pretrained LLMs trained specifically for radiology, and fewer still that have been trained in a low data setting and gone on to produce comparable results in fine-tuning tasks. We present RadLing, a continuously pretrained LLM using Electra-small (Clark et al., 2020) architecture, trained using over 500K radiology reports, that can compete with state-of-the-art results for fine tuning tasks in radiology domain. Our main contribution in this paper is knowledge-aware masking which is a taxonomic knowledge-assisted pretraining task that dynamically masks tokens to inject knowledge during pretraining. In addition, we also introduce an knowledge base-aided vocabulary extension to adapt the general tokenization vocabulary to radiology domain.
- Rikhiya Ghosh (7 papers)
- Sanjeev Kumar Karn (10 papers)
- Manuela Daniela Danu (2 papers)
- Larisa Micu (1 paper)
- Ramya Vunikili (2 papers)
- Oladimeji Farri (12 papers)