Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Ophtha-LLaMA2: A Large Language Model for Ophthalmology (2312.04906v1)

Published 8 Dec 2023 in cs.CL

Abstract: In recent years, pre-trained LLMs have achieved tremendous success in the field of NLP. Prior studies have primarily focused on general and generic domains, with relatively less research on specialized LLMs in the medical field. The specialization and high accuracy requirements for diagnosis in the medical field, as well as the challenges in collecting large-scale data, have constrained the application and development of LLMs in medical scenarios. In the field of ophthalmology, clinical diagnosis mainly relies on doctors' interpretation of reports and making diagnostic decisions. In order to take advantage of LLMs to provide decision support for doctors, we collected three modalities of ophthalmic report data and fine-tuned the LLaMA2 model, successfully constructing an LLM termed the "Ophtha-LLaMA2" specifically tailored for ophthalmic disease diagnosis. Inference test results show that even with a smaller fine-tuning dataset, Ophtha-LLaMA2 performs significantly better in ophthalmic diagnosis compared to other LLMs. It demonstrates that the Ophtha-LLaMA2 exhibits satisfying accuracy and efficiency in ophthalmic disease diagnosis, making it a valuable tool for ophthalmologists to provide improved diagnostic support for patients. This research provides a useful reference for the application of LLMs in the field of ophthalmology, while showcasing the immense potential and prospects in this domain.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (46)
  1. Length of stay prediction in neurosurgery with russian gpt-3 language model compared to human expectations. Studies in Health Technology and Informatics, 289:156–159, 2022.
  2. Leveraging weak supervision to perform named entity recognition in electronic health records progress notes to identify the ophthalmology exam. International Journal of Medical Informatics, 167:104864, 2022.
  3. Benchmarking a foundation llm on its ability to re-label structure names in accordance with the aapm tg-263 report. arXiv preprint arXiv:2310.03874, 2023.
  4. Artificial general intelligence for radiation oncology. arXiv preprint arXiv:2309.02590, 2023.
  5. Evaluating large language models for radiology natural language processing, 2023.
  6. Artificial general intelligence for medical imaging. arXiv preprint arXiv:2306.05480, 2023.
  7. Evaluating large language models on a highly-specialized topic. Radiation Oncology Physics, 2023.
  8. Matching exemplar as next sentence prediction (mensp): Zero-shot prompt learning for automatic scoring in science education. In International Conference on Artificial Intelligence in Education, pages 401–413. Springer, 2023.
  9. Artificial general intelligence (agi) for education. arXiv preprint arXiv:2304.12479, 2023.
  10. Context matters: A strategy to pre-train language model for science education. arXiv preprint arXiv:2301.12031, 2023.
  11. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730–27744, 2022.
  12. Trialling a large language model (chatgpt) in general practice with the applied knowledge test: observational study demonstrating opportunities and limitations in primary care. JMIR Medical Education, 9(1):e46599, 2023.
  13. Large language models in medicine. Nature medicine, 29(8):1930–1940, 2023.
  14. Blenderbot 3: a deployed conversational agent that continually learns to responsibly engage. arXiv preprint arXiv:2208.03188, 2022.
  15. Improving alignment of dialogue agents via targeted human judgements. arXiv preprint arXiv:2209.14375, 2022.
  16. An accurate deep learning model for clinical entity recognition from clinical notes. IEEE Journal of Biomedical and Health Informatics, 25(10):3804–3811, 2021.
  17. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
  18. Palm: Scaling language modeling with pathways. arXiv preprint arXiv:2204.02311, 2022.
  19. Gpt-3: Its nature, scope, limits, and consequences. Minds and Machines, 30:681–694, 2020.
  20. OpenAI. Gpt-4 technical report, 2023.
  21. The dawn of lmms: Preliminary explorations with gpt-4v(ision), 2023.
  22. OpenAI. Gpt-4v(ision) system card, 2023.
  23. Capabilities of gpt-4 on medical challenge problems, 2023.
  24. Performance of chatgpt as an ai-assisted decision support tool in medicine: a proof-of-concept study for interpreting symptoms and management of common cardiac conditions (amstelheart-2). medRxiv, pages 2023–03, 2023.
  25. Evaluating gpt as an adjunct for radiologic decision making: Gpt-4 versus gpt-3.5 in a breast imaging pilot. Journal of the American College of Radiology, 2023.
  26. Summary of chatgpt/gpt-4 research and perspective towards the future of large language models. arXiv preprint arXiv:2304.01852, 2023.
  27. Putting chatgpt’s medical advice to the (turing) test. medRxiv, pages 2023–01, 2023.
  28. Mededit: Model editing for medical question answering with external knowledge bases. arXiv preprint arXiv:2309.16035, 2023.
  29. Prompt engineering for healthcare: Methodologies and applications. arXiv preprint arXiv:2304.14670, 2023.
  30. Biomedgpt: A unified and generalist biomedical generative pre-trained transformer for vision, language, and multimodal tasks. arXiv preprint arXiv:2305.17100, 2023.
  31. Pharmacygpt: The ai pharmacist. arXiv preprint arXiv:2307.10432, 2023.
  32. Exploring multimodal approaches for alzheimer’s disease detection using patient speech transcript and audio data. arXiv preprint arXiv:2307.02514, 2023.
  33. Chatabl: Abductive learning via natural language interaction with chatgpt. arXiv preprint arXiv:2304.11107, 2023.
  34. Review of large vision models and visual prompt engineering. arXiv preprint arXiv:2307.00855, 2023.
  35. Ad-autogpt: An autonomous gpt for alzheimer’s disease infodemiology. arXiv preprint arXiv:2306.10095, 2023.
  36. Cohortgpt: An enhanced gpt for participant recruitment in clinical study. arXiv preprint arXiv:2307.11346, 2023.
  37. Chatradio-valuer: A chat large language model for generalizable radiology report generation based on multi-institution and multi-system data. arXiv preprint arXiv:2310.05242, 2023.
  38. Deid-gpt: Zero-shot medical text de-identification by gpt-4. arXiv preprint arXiv:2303.11032, 2023.
  39. Auggpt: Leveraging chatgpt for text data augmentation, 2023.
  40. Altered spontaneous brain activity patterns of meibomian gland dysfunction in severely obese population measured using the fractional amplitude of low-frequency fluctuations. Frontiers in Psychiatry, 13:914039, 2022.
  41. Altered brain activity in patients with diabetic retinopathy using regional homogeneity: a resting-state fmri study. Endocrine Practice, 25(4):320–327, 2019.
  42. Visualization of focal thinning of the ganglion cell–inner plexiform layer in patients with mild cognitive impairment and alzheimer’s disease. Journal of Alzheimer’s disease, 64(4):1261–1273, 2018.
  43. Retinal microvasculature alteration in active thyroid-associated ophthalmopathy. Endocrine Practice, 24(7):658–667, 2018.
  44. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288, 2023.
  45. Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685, 2021.
  46. Chin-Yew Lin. Rouge: A package for automatic evaluation of summaries. In Text summarization branches out, pages 74–81, 2004.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (17)
  1. Huan Zhao (109 papers)
  2. Qian Ling (1 paper)
  3. Yi Pan (79 papers)
  4. Tianyang Zhong (19 papers)
  5. Jin-Yu Hu (3 papers)
  6. Junjie Yao (19 papers)
  7. Fengqian Xiao (1 paper)
  8. Zhenxiang Xiao (7 papers)
  9. Yutong Zhang (34 papers)
  10. San-Hua Xu (1 paper)
  11. Shi-Nan Wu (2 papers)
  12. Min Kang (3 papers)
  13. Zihao Wu (100 papers)
  14. Zhengliang Liu (91 papers)
  15. Xi Jiang (53 papers)
  16. Tianming Liu (161 papers)
  17. Yi Shao (8 papers)
Citations (6)