Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Does the Magic of BERT Apply to Medical Code Assignment? A Quantitative Study (2103.06511v2)

Published 11 Mar 2021 in cs.CL

Abstract: Unsupervised pretraining is an integral part of many natural language processing systems, and transfer learning with LLMs has achieved remarkable results in many downstream tasks. In the clinical application of medical code assignment, diagnosis and procedure codes are inferred from lengthy clinical notes such as hospital discharge summaries. However, it is not clear if pretrained models are useful for medical code prediction without further architecture engineering. This paper conducts a comprehensive quantitative analysis of various contextualized LLMs' performance, pretrained in different domains, for medical code assignment from clinical notes. We propose a hierarchical fine-tuning architecture to capture interactions between distant words and adopt label-wise attention to exploit label information. Contrary to current trends, we demonstrate that a carefully trained classical CNN outperforms attention-based models on a MIMIC-III subset with frequent codes. Our empirical findings suggest directions for improving the medical code assignment application.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Shaoxiong Ji (39 papers)
  2. Matti Hölttä (1 paper)
  3. Pekka Marttinen (56 papers)
Citations (64)

Summary

We haven't generated a summary for this paper yet.