In the field of healthcare, the efficient extraction of information from unstructured clinical data, such as electronic health records (EHRs), plays a pivotal role in improving patient care. Specialized NLP models are deployed for such tasks; however, developing these models requires a significant amount of accurately labeled data, which is both resource-intensive and cost-prohibitive due to the need for manual annotation by medical experts.
This paper introduces an innovative method that leverages LLMs in combination with human expertise to expedite the annotation process for medical text data while maintaining high levels of accuracy. By pairing LLMs with human annotators, the process of creating labeled datasets is significantly accelerated, thus reducing the burden on human experts.
The research evaluates the effectiveness of this method through a medical information extraction task that centers on the identification and association of medication information within clinical text. Two phases of annotation are considered: Base Annotation, where initial labels are applied, and Refinement Annotation, where these labels are adjusted to ensure accuracy. They compare the performance of the LLM-assisted annotation workflow with a fully manual process using expert human annotation and refinement.
The evaluation of the findings focused on medication extraction task demonstrates that the LLM-based annotation achieves similar quality levels to the manual process while reducing annotation time by 58%. Subgroup analysis highlights the utility of LLMs even for expert annotators, with time savings of 26%, suggesting that even highly skilled annotators benefit from LLM assistance.
Moreover, the paper contributes a set of labels for medication extraction using a public medical dataset (MIMIC-IV-Note) and discusses future directions for integrating LLMs to enhance medical NLP.
In summary, this research presents a compelling use case for incorporating LLMs into the data annotation workflow, a method that shows promise in overcoming the labeling bottleneck in medical NLP tasks. As LLMs continue to improve, they may profoundly impact the field by providing efficient ways to organize and access critical information locked within unstructured clinical data.