Papers
Topics
Authors
Recent
2000 character limit reached

Text2MDT: Extracting Medical Decision Trees from Medical Texts (2401.02034v1)

Published 4 Jan 2024 in cs.CL

Abstract: Knowledge of the medical decision process, which can be modeled as medical decision trees (MDTs), is critical to build clinical decision support systems. However, the current MDT construction methods rely heavily on time-consuming and laborious manual annotation. In this work, we propose a novel task, Text2MDT, to explore the automatic extraction of MDTs from medical texts such as medical guidelines and textbooks. We normalize the form of the MDT and create an annotated Text-to-MDT dataset in Chinese with the participation of medical experts. We investigate two different methods for the Text2MDT tasks: (a) an end-to-end framework which only relies on a GPT style LLMs (LLM) instruction tuning to generate all the node information and tree structures. (b) The pipeline framework which decomposes the Text2MDT task to three subtasks. Experiments on our Text2MDT dataset demonstrate that: (a) the end-to-end method basd on LLMs (7B parameters or larger) show promising results, and successfully outperform the pipeline methods. (b) The chain-of-thought (COT) prompting method \cite{Wei2022ChainOT} can improve the performance of the fine-tuned LLMs on the Text2MDT test set. (c) the lightweight pipelined method based on encoder-based pretrained models can perform comparably with LLMs with model complexity two magnititudes smaller. Our Text2MDT dataset is open-sourced at \url{https://tianchi.aliyun.com/dataset/95414}, and the source codes are open-sourced at \url{https://github.com/michael-wzhu/text2dt}.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (80)
  1. Rule-based expert systems. Handbook of measuring system design .
  2. Predicting life expectancy with a long short-term memory recurrent neural network using electronic medical records. BMC medical informatics and decision making 19, 1–15.
  3. Aspect level sentiment analysis based on relation gated graph convolutional network. Journal of ZheJiang University (Engineering Science) 57, 437–445.
  4. A coefficient of agreement for nominal scales. Educational and Psychological Measurement 20, 37–46. arXiv:https://doi.org/10.1177/001316446002000104.
  5. CONTaiNER: Few-shot named entity recognition via contrastive learning, in: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Association for Computational Linguistics, Dublin, Ireland. pp. 6338–6353. URL: https://aclanthology.org/2022.acl-long.439, doi:10.18653/v1/2022.acl-long.439.
  6. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 .
  7. BERT: Pre-training of deep bidirectional transformers for language understanding, in: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Association for Computational Linguistics, Minneapolis, Minnesota. pp. 4171–4186. URL: https://www.aclweb.org/anthology/N19-1423, doi:10.18653/v1/N19-1423.
  8. Deep biaffine attention for neural dependency parsing. arXiv preprint arXiv:1611.01734 .
  9. Deep biaffine attention for neural dependency parsing, in: 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings, OpenReview.net. URL: https://openreview.net/forum?id=Hk95PK9le.
  10. Transition-based parsing with stack-transformers, in: Cohn, T., He, Y., Liu, Y. (Eds.), Findings of the Association for Computational Linguistics: EMNLP 2020, Association for Computational Linguistics, Online. pp. 1001–1007. URL: https://aclanthology.org/2020.findings-emnlp.89, doi:10.18653/v1/2020.findings-emnlp.89.
  11. Clinical decision support systems. John Wiley & Sons, Inc. .
  12. F-pabee: Flexible-patience-based early exiting for single-label and multi-label text classification tasks. ArXiv abs/2305.11916. URL: https://api.semanticscholar.org/CorpusID:258546016.
  13. Domain-specific language model pretraining for biomedical natural language processing. arXiv:arXiv:2007.15779.
  14. Global attention decoder for chinese spelling error correction, in: Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pp. 1419–1428.
  15. Global attention decoder for Chinese spelling error correction, in: Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, Association for Computational Linguistics, Online. pp. 1419–1428. URL: https://aclanthology.org/2021.findings-acl.122, doi:10.18653/v1/2021.findings-acl.122.
  16. Medical information extraction in the age of deep learning. Yearbook of medical informatics 29, 208–220.
  17. Degree: A data-efficient generation-based event extraction model, in: Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 1890–1908.
  18. Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685 .
  19. Medical text classification using convolutional neural networks, in: Informatics for Health: Connected Citizen-Led Wellness and Population Health. IOS Press, pp. 246–250.
  20. Medical expert systems – a study of trust and acceptance by healthcare stakeholders. Springer, Cham .
  21. The role of: A novel scientific knowledge graph representation and construction model, in: Teredesai, A., Kumar, V., Li, Y., Rosales, R., Terzi, E., Karypis, G. (Eds.), Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD 2019, Anchorage, AK, USA, August 4-8, 2019, ACM. pp. 1634–1642. URL: https://doi.org/10.1145/3292500.3330942, doi:10.1145/3292500.3330942.
  22. A survey on semantic parsing. ArXiv abs/1812.00978. URL: https://api.semanticscholar.org/CorpusID:54181011.
  23. Information extraction from electronic medical documents: state of the art and future research directions. Knowledge and Information Systems 65, 463–516.
  24. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension, in: Annual Meeting of the Association for Computational Linguistics.
  25. Sequence generation with label augmentation for relation extraction. arXiv preprint arXiv:2212.14266 .
  26. Cmmlu: Measuring massive multitask language understanding in chinese. arXiv preprint arXiv:2306.09212 .
  27. Graph neural network-based diagnosis prediction. Big Data 8, 379–390.
  28. Chip2022 shared task overview: Medical causal entity relationship extraction, in: Tang, B., Chen, Q., Lin, H., Wu, F., Liu, L., Hao, T., Wang, Y., Wang, H., Lei, J., Li, Z., Zong, H. (Eds.), Health Information Processing. Evaluation Track Papers, Springer Nature Singapore, Singapore. pp. 51–56.
  29. Fixing weight decay regularization in adam. ArXiv abs/1711.05101. URL: https://api.semanticscholar.org/CorpusID:3312944.
  30. Unified structure generation for universal information extraction, in: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Association for Computational Linguistics, Dublin, Ireland. pp. 5755–5772. URL: https://aclanthology.org/2022.acl-long.395, doi:10.18653/v1/2022.acl-long.395.
  31. Reactive, proactive, and extensible situation-awareness in ambient assisted living. Expert Systems with Applications 76, 21–35.
  32. A guided tour to approximate string matching. ACM Comput. Surv. 33, 31–88. URL: https://api.semanticscholar.org/CorpusID:207551224.
  33. Article: Medical expert system- a comprehensive review. International Journal of Computer Applications 130, 44–50. Published by Foundation of Computer Science (FCS), NY, USA.
  34. An effective neural network model for graph-based dependency parsing, in: Annual Meeting of the Association for Computational Linguistics. URL: https://api.semanticscholar.org/CorpusID:2102270.
  35. Expert systems: Definitions, advantages and issues in medical field applications. Expert Systems with Applications 177, 114900.
  36. Parsing natural scenes and natural language with recursive neural networks, in: International Conference on Machine Learning. URL: https://api.semanticscholar.org/CorpusID:18690358.
  37. Medical knowledge graph to enhance fraud, waste, and abuse detection on claim data: Model development and performance evaluation. JMIR Med Inform 8, e17653. URL: http://medinform.jmir.org/2020/7/e17653/, doi:10.2196/17653.
  38. An introduction to conditional random fields. Found. Trends Mach. Learn. 4, 267–373. URL: https://api.semanticscholar.org/CorpusID:342976.
  39. Llama 2: Open foundation and fine-tuned chat models. ArXiv abs/2307.09288. URL: https://api.semanticscholar.org/CorpusID:259950998.
  40. Automated extraction of medical expert system rules from clinical databases on rough set theory. Inf. Sci. 112, 67–84.
  41. Translating math word problem to expression tree, in: Riloff, E., Chiang, D., Hockenmaier, J., Tsujii, J. (Eds.), Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, October 31 - November 4, 2018, Association for Computational Linguistics. pp. 1064–1069. URL: https://doi.org/10.18653/v1/d18-1132, doi:10.18653/v1/d18-1132.
  42. Stanford: Probabilistic edit distance metrics for STS, in: Agirre, E., Bos, J., Diab, M., Manandhar, S., Marton, Y., Yuret, D. (Eds.), *SEM 2012: The First Joint Conference on Lexical and Computational Semantics – Volume 1: Proceedings of the main conference and the shared task, and Volume 2: Proceedings of the Sixth International Workshop on Semantic Evaluation (SemEval 2012), Association for Computational Linguistics, Montréal, Canada. pp. 648–654. URL: https://aclanthology.org/S12-1096.
  43. Unire: A unified label space for entity relation extraction, in: Zong, C., Xia, F., Li, W., Navigli, R. (Eds.), Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, ACL/IJCNLP 2021, (Volume 1: Long Papers), Virtual Event, August 1-6, 2021, Association for Computational Linguistics. pp. 220–231. URL: https://doi.org/10.18653/v1/2021.acl-long.19, doi:10.18653/v1/2021.acl-long.19.
  44. Tplinker: Single-stage joint extraction of entities and relations through token pair linking. Proceedings of the 28th International Conference on Computational Linguistics .
  45. Chain of thought prompting elicits reasoning in large language models. ArXiv abs/2201.11903. URL: https://api.semanticscholar.org/CorpusID:246411621.
  46. A novel cascade binary tagging framework for relational triple extraction, in: Jurafsky, D., Chai, J., Schluter, N., Tetreault, J.R. (Eds.), Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Online, July 5-10, 2020, Association for Computational Linguistics. pp. 1476–1488. URL: https://doi.org/10.18653/v1/2020.acl-main.136, doi:10.18653/v1/2020.acl-main.136.
  47. A goal-driven tree-structured neural model for math word problems, in: Kraus, S. (Ed.), Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI 2019, Macao, China, August 10-16, 2019, ijcai.org. pp. 5299–5305. URL: https://doi.org/10.24963/ijcai.2019/736, doi:10.24963/ijcai.2019/736.
  48. A unified generative framework for various ner subtasks. ArXiv abs/2106.01223. URL: https://api.semanticscholar.org/CorpusID:235294002.
  49. A unified generative framework for various NER subtasks, in: Zong, C., Xia, F., Li, W., Navigli, R. (Eds.), Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Association for Computational Linguistics, Online. pp. 5808–5822. URL: https://aclanthology.org/2021.acl-long.451, doi:10.18653/v1/2021.acl-long.451.
  50. Baichuan 2: Open large-scale language models. ArXiv abs/2309.10305. URL: https://api.semanticscholar.org/CorpusID:261951743.
  51. A survey of information extraction based on deep learning. Applied Sciences 12, 9691.
  52. Named entity recognition as dependency parsing, in: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics, Online. pp. 6470–6476. URL: https://aclanthology.org/2020.acl-main.577, doi:10.18653/v1/2020.acl-main.577.
  53. Syntaxsqlnet: Syntax tree networks for complex and cross-domain text-to-sql task, in: Riloff, E., Chiang, D., Hockenmaier, J., Tsujii, J. (Eds.), Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, October 31 - November 4, 2018, Association for Computational Linguistics. pp. 1653–1663. URL: https://doi.org/10.18653/v1/d18-1193, doi:10.18653/v1/d18-1193.
  54. A survey of syntactic-semantic parsing based on constituent and dependency structures. Science China Technological Sciences 63, 1898 – 1920. URL: https://api.semanticscholar.org/CorpusID:219956190.
  55. CBLUE: A Chinese biomedical language understanding evaluation benchmark, in: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Association for Computational Linguistics, Dublin, Ireland. pp. 7888–7915. URL: https://aclanthology.org/2022.acl-long.544, doi:10.18653/v1/2022.acl-long.544.
  56. De-bias for generative extraction in unified ner task, in: Annual Meeting of the Association for Computational Linguistics.
  57. De-bias for generative extraction in unified NER task, in: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Association for Computational Linguistics, Dublin, Ireland. pp. 808–818. URL: https://aclanthology.org/2022.acl-long.59, doi:10.18653/v1/2022.acl-long.59.
  58. Multi-view reasoning: Consistent contrastive learning for math word problem, in: Conference on Empirical Methods in Natural Language Processing. URL: https://api.semanticscholar.org/CorpusID:253080836.
  59. Transition-based parsing for deep dependency structures. Computational Linguistics 42, 353–389. URL: https://aclanthology.org/J16-3001, doi:10.1162/COLI_a_00252.
  60. Nag-ner: a unified non-autoregressive generation framework for various ner tasks, in: Annual Meeting of the Association for Computational Linguistics. URL: https://api.semanticscholar.org/CorpusID:259370837.
  61. Fastner: Speeding up inferences for named entity recognition tasks, in: International Conference on Advanced Data Mining and Applications. URL: https://api.semanticscholar.org/CorpusID:265214231.
  62. Automatic student network search for knowledge distillation. 2020 25th International Conference on Pattern Recognition (ICPR) , 2446–2453.
  63. A survey of large language models. ArXiv abs/2303.18223. URL: https://api.semanticscholar.org/CorpusID:257900969.
  64. Automatic model selection with large language models for reasoning, in: Conference on Empirical Methods in Natural Language Processing. URL: https://api.semanticscholar.org/CorpusID:258840882.
  65. Seq2sql: Generating structured queries from natural language using reinforcement learning. CoRR abs/1709.00103. arXiv:1709.00103.
  66. Natural Language Processing for Smart Healthcare. arXiv e-prints , arXiv:2110.15803doi:10.48550/arXiv.2110.15803, arXiv:2110.15803.
  67. Analysis of the health information needs of diabetics in china. Studies in health technology and informatics 264, 487–491. URL: https://api.semanticscholar.org/CorpusID:201617388.
  68. AutoRC: Improving BERT based relation classification models via architecture search, in: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing: Student Research Workshop, Association for Computational Linguistics, Online. pp. 33–43. URL: https://aclanthology.org/2021.acl-srw.4, doi:10.18653/v1/2021.acl-srw.4.
  69. MVP-BERT: Multi-vocab pre-training for Chinese BERT, in: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing: Student Research Workshop, Association for Computational Linguistics, Online. pp. 260–269. URL: https://aclanthology.org/2021.acl-srw.27, doi:10.18653/v1/2021.acl-srw.27.
  70. Mvp-bert: Multi-vocab pre-training for chinese bert, in: Annual Meeting of the Association for Computational Linguistics. URL: https://api.semanticscholar.org/CorpusID:237331564.
  71. paht_nlp @ mediqa 2021: Multi-grained query focused multi-answer summarization, in: Workshop on Biomedical Natural Language Processing. URL: https://api.semanticscholar.org/CorpusID:235097590.
  72. Extracting decision trees from medical texts: An overview of the text2dt track in chip2022, in: Tang, B., Chen, Q., Lin, H., Wu, F., Liu, L., Hao, T., Wang, Y., Wang, H., Lei, J., Li, Z., Zong, H. (Eds.), Health Information Processing. Evaluation Track Papers, Springer Nature Singapore, Singapore. pp. 89–102.
  73. Discovering better model architectures for medical query understanding, in: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Industry Papers, Association for Computational Linguistics, Online. pp. 230–237. URL: https://aclanthology.org/2021.naacl-industry.29, doi:10.18653/v1/2021.naacl-industry.29.
  74. AutoRC: Improving BERT Based Relation Classification Models via Architecture Search. arXiv e-prints , arXiv:2009.10680doi:10.48550/arXiv.2009.10680, arXiv:2009.10680.
  75. SPT: Learning to selectively insert prompts for better prompt tuning, in: Bouamor, H., Pino, J., Bali, K. (Eds.), Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics, Singapore. pp. 11862–11878. URL: https://aclanthology.org/2023.emnlp-main.727.
  76. Acf: Aligned contrastive finetuning for language and vision tasks. ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) , 1–5URL: https://api.semanticscholar.org/CorpusID:258533119.
  77. Chatmed: A chinese medical large language model. https://github.com/michael-wzhu/ChatMed.
  78. Autotrans: Automating transformer design via reinforced architecture search, in: Wang, L., Feng, Y., Hong, Y., He, R. (Eds.), Natural Language Processing and Chinese Computing, Springer International Publishing, Cham. pp. 169–182.
  79. PromptCBLUE: A Chinese Prompt Tuning Benchmark for the Medical Domain. arXiv e-prints , arXiv:2310.14151doi:10.48550/arXiv.2310.14151, arXiv:2310.14151.
  80. PANLP at MEDIQA 2019: Pre-trained language models, transfer learning and knowledge distillation, in: Proceedings of the 18th BioNLP Workshop and Shared Task, Association for Computational Linguistics, Florence, Italy. pp. 380–388. URL: https://aclanthology.org/W19-5040, doi:10.18653/v1/W19-5040.
Citations (4)

Summary

We haven't generated a summary for this paper yet.

Whiteboard

Video Overview

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.