Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Contextualized Medication Information Extraction Using Transformer-based Deep Learning Architectures (2303.08259v1)

Published 14 Mar 2023 in cs.CL and cs.AI

Abstract: Objective: To develop a NLP system to extract medications and contextual information that help understand drug changes. This project is part of the 2022 n2c2 challenge. Materials and methods: We developed NLP systems for medication mention extraction, event classification (indicating medication changes discussed or not), and context classification to classify medication changes context into 5 orthogonal dimensions related to drug changes. We explored 6 state-of-the-art pretrained transformer models for the three subtasks, including GatorTron, a LLM pretrained using >90 billion words of text (including >80 billion words from >290 million clinical notes identified at the University of Florida Health). We evaluated our NLP systems using annotated data and evaluation scripts provided by the 2022 n2c2 organizers. Results:Our GatorTron models achieved the best F1-scores of 0.9828 for medication extraction (ranked 3rd), 0.9379 for event classification (ranked 2nd), and the best micro-average accuracy of 0.9126 for context classification. GatorTron outperformed existing transformer models pretrained using smaller general English text and clinical text corpora, indicating the advantage of LLMs. Conclusion: This study demonstrated the advantage of using large transformer models for contextual medication information extraction from clinical narratives.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Aokun Chen (12 papers)
  2. Zehao Yu (41 papers)
  3. Xi Yang (160 papers)
  4. Yi Guo (115 papers)
  5. Jiang Bian (229 papers)
  6. Yonghui Wu (115 papers)
Citations (16)