Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Survey of Knowledge-Intensive NLP with Pre-Trained Language Models (2202.08772v1)

Published 17 Feb 2022 in cs.CL

Abstract: With the increasing of model capacity brought by pre-trained LLMs, there emerges boosting needs for more knowledgeable NLP models with advanced functionalities including providing and making flexible use of encyclopedic and commonsense knowledge. The mere pre-trained LLMs, however, lack the capacity of handling such knowledge-intensive NLP tasks alone. To address this challenge, large numbers of pre-trained LLMs augmented with external knowledge sources are proposed and in rapid development. In this paper, we aim to summarize the current progress of pre-trained LLM-based knowledge-enhanced models (PLMKEs) by dissecting their three vital elements: knowledge sources, knowledge-intensive NLP tasks, and knowledge fusion methods. Finally, we present the challenges of PLMKEs based on the discussion regarding the three elements and attempt to provide NLP practitioners with potential directions for further research.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Da Yin (35 papers)
  2. Li Dong (154 papers)
  3. Hao Cheng (190 papers)
  4. Xiaodong Liu (162 papers)
  5. Kai-Wei Chang (292 papers)
  6. Furu Wei (291 papers)
  7. Jianfeng Gao (344 papers)
Citations (31)