Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

PediatricsGPT: Large Language Models as Chinese Medical Assistants for Pediatric Applications (2405.19266v4)

Published 29 May 2024 in cs.CL

Abstract: Developing intelligent pediatric consultation systems offers promising prospects for improving diagnostic efficiency, especially in China, where healthcare resources are scarce. Despite recent advances in LLMs for Chinese medicine, their performance is sub-optimal in pediatric applications due to inadequate instruction data and vulnerable training procedures. To address the above issues, this paper builds PedCorpus, a high-quality dataset of over 300,000 multi-task instructions from pediatric textbooks, guidelines, and knowledge graph resources to fulfil diverse diagnostic demands. Upon well-designed PedCorpus, we propose PediatricsGPT, the first Chinese pediatric LLM assistant built on a systematic and robust training pipeline. In the continuous pre-training phase, we introduce a hybrid instruction pre-training mechanism to mitigate the internal-injected knowledge inconsistency of LLMs for medical domain adaptation. Immediately, the full-parameter Supervised Fine-Tuning (SFT) is utilized to incorporate the general medical knowledge schema into the models. After that, we devise a direct following preference optimization to enhance the generation of pediatrician-like humanistic responses. In the parameter-efficient secondary SFT phase, a mixture of universal-specific experts strategy is presented to resolve the competency conflict between medical generalist and pediatric expertise mastery. Extensive results based on the metrics, GPT-4, and doctor evaluations on distinct doctor downstream tasks show that PediatricsGPT consistently outperforms previous Chinese medical LLMs. Our model and dataset will be open-source for community development.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (14)
  1. Dingkang Yang (57 papers)
  2. Jinjie Wei (11 papers)
  3. Dongling Xiao (10 papers)
  4. Shunli Wang (21 papers)
  5. Tong Wu (228 papers)
  6. Gang Li (579 papers)
  7. Mingcheng Li (25 papers)
  8. Shuaibing Wang (9 papers)
  9. Jiawei Chen (160 papers)
  10. Yue Jiang (104 papers)
  11. Qingyao Xu (9 papers)
  12. Ke Li (722 papers)
  13. Peng Zhai (16 papers)
  14. Lihua Zhang (68 papers)
Citations (6)