Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

KEHRL: Learning Knowledge-Enhanced Language Representations with Hierarchical Reinforcement Learning (2406.16374v1)

Published 24 Jun 2024 in cs.CL

Abstract: Knowledge-enhanced pre-trained LLMs (KEPLMs) leverage relation triples from knowledge graphs (KGs) and integrate these external data sources into LLMs via self-supervised learning. Previous works treat knowledge enhancement as two independent operations, i.e., knowledge injection and knowledge integration. In this paper, we propose to learn Knowledge-Enhanced language representations with Hierarchical Reinforcement Learning (KEHRL), which jointly addresses the problems of detecting positions for knowledge injection and integrating external knowledge into the model in order to avoid injecting inaccurate or irrelevant knowledge. Specifically, a high-level reinforcement learning (RL) agent utilizes both internal and prior knowledge to iteratively detect essential positions in texts for knowledge injection, which filters out less meaningful entities to avoid diverting the knowledge learning direction. Once the entity positions are selected, a relevant triple filtration module is triggered to perform low-level RL to dynamically refine the triples associated with polysemic entities through binary-valued actions. Experiments validate KEHRL's effectiveness in probing factual knowledge and enhancing the model's performance on various natural language understanding tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Dongyang Li (41 papers)
  2. Taolin Zhang (34 papers)
  3. Longtao Huang (27 papers)
  4. Chengyu Wang (93 papers)
  5. Xiaofeng He (33 papers)
  6. Hui Xue (109 papers)