Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

JAKET: Joint Pre-training of Knowledge Graph and Language Understanding (2010.00796v1)

Published 2 Oct 2020 in cs.CL

Abstract: Knowledge graphs (KGs) contain rich information about world knowledge, entities and relations. Thus, they can be great supplements to existing pre-trained LLMs. However, it remains a challenge to efficiently integrate information from KG into LLMing. And the understanding of a knowledge graph requires related context. We propose a novel joint pre-training framework, JAKET, to model both the knowledge graph and language. The knowledge module and language module provide essential information to mutually assist each other: the knowledge module produces embeddings for entities in text while the language module generates context-aware initial embeddings for entities and relations in the graph. Our design enables the pre-trained model to easily adapt to unseen knowledge graphs in new domains. Experimental results on several knowledge-aware NLP tasks show that our proposed framework achieves superior performance by effectively leveraging knowledge in language understanding.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Donghan Yu (18 papers)
  2. Chenguang Zhu (100 papers)
  3. Yiming Yang (151 papers)
  4. Michael Zeng (76 papers)
Citations (131)