Editing Factual Knowledge and Explanatory Ability of Medical Large Language Models (2402.18099v3)
Abstract: Model editing aims to precisely alter the behaviors of LLMs in relation to specific knowledge, while leaving unrelated knowledge intact. This approach has proven effective in addressing issues of hallucination and outdated information in LLMs. However, the potential of using model editing to modify knowledge in the medical field remains largely unexplored, even though resolving hallucination is a pressing need in this area. Our observations indicate that current methods face significant challenges in dealing with specialized and complex knowledge in medical domain. Therefore, we propose MedLaSA, a novel Layer-wise Scalable Adapter strategy for medical model editing. MedLaSA harnesses the strengths of both adding extra parameters and locate-then-edit methods for medical model editing. We utilize causal tracing to identify the association of knowledge in neurons across different layers, and generate a corresponding scale set from the association value for each piece of knowledge. Subsequently, we incorporate scalable adapters into the dense layers of LLMs. These adapters are assigned scaling values based on the corresponding specific knowledge, which allows for the adjustment of the adapter's weight and rank. The more similar the content, the more consistent the scale between them. This ensures precise editing of semantically identical knowledge while avoiding impact on unrelated knowledge. To evaluate the editing impact on the behaviours of LLMs, we propose two model editing studies for medical domain: (1) editing factual knowledge for medical specialization and (2) editing the explanatory ability for complex knowledge. We build two novel medical benchmarking datasets and introduce a series of challenging and comprehensive metrics. Extensive experiments on medical LLMs demonstrate the editing efficiency of MedLaSA, without affecting unrelated knowledge.
- MEDITRON-70B: Scaling Medical Pretraining for Large Language Models.
- Knowledge Neurons in Pretrained Transformers. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 8493–8502, Dublin, Ireland. Association for Computational Linguistics.
- Editing Factual Knowledge in Language Models. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 6491–6506.
- Calibrating Factual Knowledge in Pretrained Language Models. arXiv preprint arXiv:2210.03329.
- Trends in Integration of Knowledge and Large Language Models: A Survey and Taxonomy of Methods, Benchmarks, and Applications. arXiv preprint arXiv:2311.05876.
- Transformer Feed-Forward Layers Build Predictions by Promoting Concepts in the Vocabulary Space. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 30–45.
- Transformer Feed-Forward Layers Are Key-Value Memories. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 5484–5495.
- Aging with GRACE: Lifelong Model Editing with Discrete Key-Value Adaptors. In NeurIPS 2022 Workshop on Robustness in Sequence Modeling.
- LoRA: Low-Rank Adaptation of Large Language Models. arXiv preprint arXiv:2106.09685.
- Transformer-Patcher: One Mistake Worth One Neuron. In The Eleventh International Conference on Learning Representations.
- DRKG - Drug Repurposing Knowledge Graph for Covid-19. https://github.com/gnn4dr/DRKG/.
- Mert Karabacak and Konstantinos Margetis. 2023. Embracing Large Language Models for Medical Applications: Opportunities and Challenges. Cureus, 15(5).
- Biobert: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics, 36(4):1234–1240.
- PMET: Precise Model Editing in a Transformer. arXiv preprint arXiv:2308.08742.
- Locating and Editing Factual Associations in GPT. Advances in Neural Information Processing Systems, 35:17359–17372.
- Mass-Editing Memory in a Transformer. arXiv preprint arXiv:2210.07229.
- Fast Model Editing at Scale. In International Conference on Learning Representations.
- Memory-based model editing at scale. In International Conference on Machine Learning, pages 15817–15831. PMLR.
- OpenAI. 2023. GPT-4 Technical Report. ArXiv, abs/2303.08774.
- MedMCQA: A Large-scale Multi-subject Multi-choice Dataset for Medical domain Question Answering. In Proceedings of the Conference on Health, Inference, and Learning, volume 174 of Proceedings of Machine Learning Research, pages 248–260. PMLR.
- Judea Pearl. 2022. Direct and indirect effects. In Probabilistic and causal inference: the works of Judea Pearl, pages 373–392.
- Language Models as Knowledge Bases? In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 2463–2473.
- Rotate: Knowledge graph embedding by relational rotation in complex space. In International Conference on Learning Representations.
- Llama 2: Open Foundation and Fine-Tuned Chat Models. arXiv preprint arXiv:2307.09288.
- Investigating Gender Bias in Language Models Using Causal Mediation Analysis. In Advances in Neural Information Processing Systems, volume 33, pages 12388–12401. Curran Associates, Inc.
- EasyEdit: An Easy-to-use Knowledge Editing Framework for Large Language Models. arXiv preprint arXiv:2308.07269.
- Knowledge Editing for Large Language Models: A Survey.
- DEPN: Detecting and Editing Privacy Neurons in Pretrained Language Models. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 2875–2886.
- Multimodal Biological Knowledge Graph Completion via Triple Co-attention Mechanism. In 2023 IEEE 39th International Conference on Data Engineering (ICDE), pages 3928–3941. IEEE.
- Editing Large Language Models: Problems, Methods, and Opportunities. arXiv preprint arXiv:2305.13172.
- MELO: Enhancing Model Editing with Neuron-Indexed Dynamic LoRA. arXiv preprint arXiv:2312.11795.
- ChatDoctor: A Medical Chat Model Fine-Tuned on a Large Language Model Meta-AI (LLaMA) Using Medical Domain Knowledge. arXiv preprint arXiv:2303.14070.
- A Comprehensive Study of Knowledge Editing for Large Language Models. arXiv preprint arXiv:2401.01286.
- Can We Edit Factual Knowledge by In-Context Learning? In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 4862–4876, Singapore. Association for Computational Linguistics.
- A survey of large language models in medicine: Progress, application, and challenge. arXiv preprint arXiv:2311.05112.
- Derong Xu (26 papers)
- Ziheng Zhang (43 papers)
- Zhihong Zhu (45 papers)
- Zhenxi Lin (11 papers)
- Qidong Liu (36 papers)
- Xian Wu (139 papers)
- Tong Xu (113 papers)
- Xiangyu Zhao (192 papers)
- Yefeng Zheng (197 papers)
- Enhong Chen (242 papers)
- Wanyu Wang (26 papers)
- Yuyang Ye (16 papers)