Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Clinical Concept Extraction for Document-Level Coding (1906.03380v1)

Published 8 Jun 2019 in cs.CL

Abstract: The text of clinical notes can be a valuable source of patient information and clinical assessments. Historically, the primary approach for exploiting clinical notes has been information extraction: linking spans of text to concepts in a detailed domain ontology. However, recent work has demonstrated the potential of supervised machine learning to extract document-level codes directly from the raw text of clinical notes. We propose to bridge the gap between the two approaches with two novel syntheses: (1) treating extracted concepts as features, which are used to supplement or replace the text of the note; (2) treating extracted concepts as labels, which are used to learn a better representation of the text. Unfortunately, the resulting concepts do not yield performance gains on the document-level clinical coding task. We explore possible explanations and future research directions.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Sarah Wiegreffe (20 papers)
  2. Edward Choi (90 papers)
  3. Sherry Yan (2 papers)
  4. Jimeng Sun (181 papers)
  5. Jacob Eisenstein (73 papers)
Citations (11)

Summary

We haven't generated a summary for this paper yet.