Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
72 tokens/sec
GPT-4o
61 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Inspire the Large Language Model by External Knowledge on BioMedical Named Entity Recognition (2309.12278v1)

Published 21 Sep 2023 in cs.CL

Abstract: LLMs have demonstrated dominating performance in many NLP tasks, especially on generative tasks. However, they often fall short in some information extraction tasks, particularly those requiring domain-specific knowledge, such as Biomedical Named Entity Recognition (NER). In this paper, inspired by Chain-of-thought, we leverage the LLM to solve the Biomedical NER step-by-step: break down the NER task into entity span extraction and entity type determination. Additionally, for entity type determination, we inject entity knowledge to address the problem that LLM's lack of domain knowledge when predicting entity category. Experimental results show a significant improvement in our two-step BioNER approach compared to previous few-shot LLM baseline. Additionally, the incorporation of external knowledge significantly enhances entity category determination performance.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (19)
  1. Biomedical entity recognition by detection and matching. arXiv preprint arXiv:2306.15736.
  2. Olivier Bodenreider. 2004. The unified medical language system (umls): integrating biomedical terminology. volume 32, pages D267–D270. Oxford University Press.
  3. The colorado richly annotated full text (craft) corpus: Multi-model annotation in the biomedical domain. Handbook of Linguistic annotation, pages 1379–1394.
  4. An activity phosphorylating tyrosine in polyoma t antigen immunoprecipitates. Cell, 18(4):925–933.
  5. Jacob Devlin Ming-Wei Chang Kenton and Lee Kristina Toutanova. 2019. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of NAACL-HLT, pages 4171–4186.
  6. Biobert: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics, 36(4):1234–1240.
  7. A unified mrc framework for named entity recognition. arXiv preprint arXiv:1910.11476.
  8. Chain of knowledge: A framework for grounding large language models with structured knowledge bases. arXiv preprint arXiv:2305.13269.
  9. Unified structure generation for universal information extraction. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 5755–5772.
  10. Aioner: all-in-one scheme-based biomedical named entity recognition using deep learning. Bioinformatics, 39(5):btad310.
  11. Xuezhe Ma and Eduard Hovy. 2016. End-to-end sequence labeling via bi-directional lstm-cnns-crf. arXiv preprint arXiv:1603.01354.
  12. Gpt-ner: Named entity recognition via large language models. arXiv preprint arXiv:2304.10428.
  13. Cross-type biomedical named entity recognition with deep multi-task learning. Bioinformatics, 35(10):1745–1752.
  14. Huner: improving biomedical ner with pretraining. Bioinformatics, 36(1):295–302.
  15. Hunflair: an easy-to-use tool for state-of-the-art biomedical named entity recognition. Bioinformatics, 37(17):2792–2794.
  16. Pubtator central: automated concept annotation for biomedical full text articles. Nucleic acids research, 47(W1):W587–W593.
  17. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35:24824–24837.
  18. Verify-and-edit: A knowledge-enhanced chain-of-thought framework. arXiv preprint arXiv:2305.03268.
  19. Universalner: Targeted distillation from large language models for open named entity recognition.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Junyi Bian (6 papers)
  2. Jiaxuan Zheng (2 papers)
  3. Yuyi Zhang (9 papers)
  4. Shanfeng Zhu (9 papers)
Citations (7)