Papers
Topics
Authors
Recent
2000 character limit reached

Document Understanding for Healthcare Referrals (2309.13184v1)

Published 22 Sep 2023 in cs.CL and cs.IR

Abstract: Reliance on scanned documents and fax communication for healthcare referrals leads to high administrative costs and errors that may affect patient care. In this work we propose a hybrid model leveraging LayoutLMv3 along with domain-specific rules to identify key patient, physician, and exam-related entities in faxed referral documents. We explore some of the challenges in applying a document understanding model to referrals, which have formats varying by medical practice, and evaluate model performance using MUC-5 metrics to obtain appropriate metrics for the practical use case. Our analysis shows the addition of domain-specific rules to the transformer model yields greatly increased precision and F1 scores, suggesting a hybrid model trained on a curated dataset can increase efficiency in referral management.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (12)
  1. C. Brown, “Health care clings to faxes as U.S. pushes electronic records,” https://news.bloomberglaw.com/health-law-and-business/health-care-clings-to-faxes-as-u-s-pushes-electronic-records, Nov. 2021, accessed: 2023-1-29.
  2. Y. Huang, T. Lv, L. Cui, Y. Lu, and F. Wei, “LayoutLMv3: Pre-training for document AI with unified text and image masking,” Apr. 2022.
  3. Y. Xu, M. Li, L. Cui, S. Huang, F. Wei, and M. Zhou, “LayoutLM: Pre-training of text and layout for document image understanding,” Dec. 2019.
  4. J. Gu, J. Kuen, V. I. Morariu, H. Zhao, N. Barmpalios, R. Jain, A. Nenkova, and T. Sun, “Unified pretraining framework for document understanding,” Apr. 2022.
  5. C. Li, B. Bi, M. Yan, W. Wang, S. Huang, F. Huang, and L. Si, “StructuralLM: Structural pre-training for form understanding,” May 2021.
  6. S. Appalaraju, B. Jasani, B. U. Kota, Y. Xie, and R. Manmatha, “DocFormer: End-to-End transformer for document understanding,” Jun. 2021.
  7. Z. Zhang, J. Ma, J. Du, L. Wang, and J. Zhang, “Multimodal pre-training based on graph attention network for document understanding,” IEEE Trans. Multimedia, pp. 1–13, 2022.
  8. Amazon Web Services, “Amazon textract,” https://aws.amazon.com/textract/, accessed: 2022-09-01.
  9. Z. Shen, R. Zhang, M. Dell, B. C. G. Lee, J. Carlson, and W. Li, “LayoutParser: A unified toolkit for deep learning based document image analysis,” Mar. 2021.
  10. M. Ester, H.-P. Kriegel, J. Sander, and X. Xu, “A density-based algorithm for discovering clusters in large spatial databases with noise,” in Proceedings of the Second International Conference on Knowledge Discovery and Data Mining.   AAAI Press, Aug. 1996, pp. 226–231.
  11. “microsoft/layoutlmv3-large,” https://huggingface.co/microsoft/layoutlmv3-large, accessed 2023-01-22.
  12. N. Chinchor and B. Sundheim, “MUC-5 evaluation metrics,” in Fifth Message Understanding Conference (MUC-5): Proceedings of a Conference Held in Baltimore, Maryland, August 25-27, 1993, 1993.

Summary

We haven't generated a summary for this paper yet.

Whiteboard

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.