Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
122 tokens/sec
GPT-4o
11 tokens/sec
Gemini 2.5 Pro Pro
47 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
4 tokens/sec
DeepSeek R1 via Azure Pro
33 tokens/sec
2000 character limit reached

Zero-shot and Few-shot Generation Strategies for Artificial Clinical Records (2403.08664v2)

Published 13 Mar 2024 in cs.CL and cs.LG

Abstract: The challenge of accessing historical patient data for clinical research, while adhering to privacy regulations, is a significant obstacle in medical science. An innovative approach to circumvent this issue involves utilising synthetic medical records that mirror real patient data without compromising individual privacy. The creation of these synthetic datasets, particularly without using actual patient data to train LLMs, presents a novel solution as gaining access to sensitive patient information to train models is also a challenge. This study assesses the capability of the Llama 2 LLM to create synthetic medical records that accurately reflect real patient information, employing zero-shot and few-shot prompting strategies for comparison against fine-tuned methodologies that do require sensitive patient data during training. We focus on generating synthetic narratives for the History of Present Illness section, utilising data from the MIMIC-IV dataset for comparison. In this work introduce a novel prompting technique that leverages a chain-of-thought approach, enhancing the model's ability to generate more accurate and contextually relevant medical narratives without prior fine-tuning. Our findings suggest that this chain-of-thought prompted approach allows the zero-shot model to achieve results on par with those of fine-tuned models, based on Rouge metrics evaluation.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (30)
  1. Optuna: A next-generation hyperparameter optimization framework. In Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining. 2623–2631.
  2. Exploring transformer text generation for medical dataset augmentation. In Proceedings of the Twelfth Language Resources and Evaluation Conference. European Language Resources Association, 4699–4708.
  3. Language Models are Few-Shot Learners. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual.
  4. Electronic health records: new opportunities for clinical research. Journal of internal medicine 274, 6 (2013), 547–560.
  5. Electronic health records to facilitate clinical research. Clinical Research in Cardiology 106 (2017), 1–9.
  6. QLoRA: Efficient Finetuning of Quantized LLMs. CoRR abs/2305.14314 (2023).
  7. Summarizing Patients’ Problems from Hospital Progress Notes Using Pre-trained Sequence-to-Sequence Models. In Proceedings of the 29th International Conference on Computational Linguistics, COLING 2022. 2979–2991.
  8. Vince Hartman and Thomas R Campion. 2022. A Day-to-Day Approach for Automating the Hospital Course Section of the Discharge Summary. In AMIA Annual Symposium Proceedings, Vol. 2022. 216.
  9. Alexander Hoerbst and Elske Ammenwerth. 2010. Electronic health records. Methods of information in medicine 49, 04 (2010), 320–336.
  10. LoRA: Low-Rank Adaptation of Large Language Models. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net.
  11. Julia Ive. 2022. Leveraging the potential of synthetic text for AI in mental healthcare. Frontiers Digit. Health 4 (2022).
  12. Generation and evaluation of artificial mental health records for Natural Language Processing. npj Digit. Medicine (2020).
  13. MIMIC-IV, a freely accessible electronic health record dataset. Scientific data 10 (2023), 1.
  14. MIMIC-III, a freely accessible critical care database. Scientific data 3 (2016), 1–9.
  15. Are synthetic clinical notes useful for real natural language processing tasks: A case study on clinical entity recognition. J. Am. Medical Informatics Assoc. 28, 10 (2021), 2193–2201. https://doi.org/10.1093/JAMIA/OCAB112
  16. Chin-Yew Lin. 2004. ROUGE: A Package for Automatic Evaluation of Summaries. In Text Summarization Branches Out. Association for Computational Linguistics, 74–81.
  17. Textual data augmentation for patient outcomes prediction. In 2021 IEEE International Conference on Bioinformatics and Biomedicine (BIBM). 2817–2821.
  18. BioGPT: generative pre-trained transformer for biomedical text generation and mining. Briefings in Bioinformatics 23, 6 (2022), bbac409.
  19. Oren Melamud and Chaitanya Shivade. 2019. Towards Automatic Generation of Shareable Synthetic Clinical Notes Using Neural Language Models. CoRR abs/1905.07002 (2019).
  20. Privacy of clinical research subjects: an integrative literature review. Journal of Empirical Research on Human Research Ethics 14, 1 (2019), 33–48.
  21. Neural Summarization of Electronic Health Records. CoRR abs/2305.15222 (2023). https://doi.org/10.48550/ARXIV.2305.15222
  22. Bleu: a Method for Automatic Evaluation of Machine Translation. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, July 6-12, 2002, Philadelphia, PA, USA. ACL, 311–318. https://doi.org/10.3115/1073083.1073135
  23. Improving language understanding by generative pre-training. (2018).
  24. Language models are unsupervised multitask learners. OpenAI blog 1, 8 (2019), 9.
  25. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. J. Mach. Learn. Res. 21 (2020), 140:1–140:67.
  26. Toward causal representation learning. Proc. IEEE 109, 5 (2021), 612–634.
  27. Llama 2: Open Foundation and Fine-Tuned Chat Models. CoRR abs/2307.09288 (2023).
  28. Attention is All you Need. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017,. 5998–6008.
  29. ColBERT-PRF: Semantic pseudo-relevance feedback for dense passage and document retrieval. ACM Transactions on the Web 17, 1 (2023), 1–39.
  30. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems 35 (2022), 24824–24837.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.