Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

KgPLM: Knowledge-guided Language Model Pre-training via Generative and Discriminative Learning (2012.03551v1)

Published 7 Dec 2020 in cs.CL and cs.AI

Abstract: Recent studies on pre-trained LLMs have demonstrated their ability to capture factual knowledge and applications in knowledge-aware downstream tasks. In this work, we present a LLM pre-training framework guided by factual knowledge completion and verification, and use the generative and discriminative approaches cooperatively to learn the model. Particularly, we investigate two learning schemes, named two-tower scheme and pipeline scheme, in training the generator and discriminator with shared parameter. Experimental results on LAMA, a set of zero-shot cloze-style question answering tasks, show that our model contains richer factual knowledge than the conventional pre-trained LLMs. Furthermore, when fine-tuned and evaluated on the MRQA shared tasks which consists of several machine reading comprehension datasets, our model achieves the state-of-the-art performance, and gains large improvements on NewsQA (+1.26 F1) and TriviaQA (+1.56 F1) over RoBERTa.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Bin He (58 papers)
  2. Xin Jiang (242 papers)
  3. Jinghui Xiao (9 papers)
  4. Qun Liu (230 papers)
Citations (18)