Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

ERICA: Improving Entity and Relation Understanding for Pre-trained Language Models via Contrastive Learning (2012.15022v2)

Published 30 Dec 2020 in cs.CL and cs.AI

Abstract: Pre-trained LLMs (PLMs) have shown superior performance on various downstream NLP tasks. However, conventional pre-training objectives do not explicitly model relational facts in text, which are crucial for textual understanding. To address this issue, we propose a novel contrastive learning framework ERICA to obtain a deep understanding of the entities and their relations in text. Specifically, we define two novel pre-training tasks to better understand entities and relations: (1) the entity discrimination task to distinguish which tail entity can be inferred by the given head entity and relation; (2) the relation discrimination task to distinguish whether two relations are close or not semantically, which involves complex relational reasoning. Experimental results demonstrate that ERICA can improve typical PLMs (BERT and RoBERTa) on several language understanding tasks, including relation extraction, entity typing and question answering, especially under low-resource settings.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Yujia Qin (41 papers)
  2. Yankai Lin (125 papers)
  3. Ryuichi Takanobu (17 papers)
  4. Zhiyuan Liu (433 papers)
  5. Peng Li (390 papers)
  6. Heng Ji (266 papers)
  7. Minlie Huang (226 papers)
  8. Maosong Sun (337 papers)
  9. Jie Zhou (687 papers)
Citations (121)

Summary

We haven't generated a summary for this paper yet.