Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Survey on Knowledge-Enhanced Pre-trained Language Models (2212.13428v1)

Published 27 Dec 2022 in cs.CL

Abstract: NLP has been revolutionized by the use of Pre-trained LLMs (PLMs) such as BERT. Despite setting new records in nearly every NLP task, PLMs still face a number of challenges including poor interpretability, weak reasoning capability, and the need for a lot of expensive annotated data when applied to downstream tasks. By integrating external knowledge into PLMs, \textit{\underline{K}nowledge-\underline{E}nhanced \underline{P}re-trained \underline{L}anguage \underline{M}odels} (KEPLMs) have the potential to overcome the above-mentioned limitations. In this paper, we examine KEPLMs systematically through a series of studies. Specifically, we outline the common types and different formats of knowledge to be integrated into KEPLMs, detail the existing methods for building and evaluating KEPLMS, present the applications of KEPLMs in downstream tasks, and discuss the future research directions. Researchers will benefit from this survey by gaining a quick and comprehensive overview of the latest developments in this field.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Chaoqi Zhen (1 paper)
  2. Yanlei Shang (3 papers)
  3. Xiangyu Liu (47 papers)
  4. Yifei Li (92 papers)
  5. Yong Chen (299 papers)
  6. Dell Zhang (26 papers)
Citations (3)