Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learn to Not Link: Exploring NIL Prediction in Entity Linking (2305.15725v1)

Published 25 May 2023 in cs.CL

Abstract: Entity linking models have achieved significant success via utilizing pretrained LLMs to capture semantic features. However, the NIL prediction problem, which aims to identify mentions without a corresponding entity in the knowledge base, has received insufficient attention. We categorize mentions linking to NIL into Missing Entity and Non-Entity Phrase, and propose an entity linking dataset NEL that focuses on the NIL prediction problem. NEL takes ambiguous entities as seeds, collects relevant mention context in the Wikipedia corpus, and ensures the presence of mentions linking to NIL by human annotation and entity masking. We conduct a series of experiments with the widely used bi-encoder and cross-encoder entity linking models, results show that both types of NIL mentions in training data have a significant influence on the accuracy of NIL prediction. Our code and dataset can be accessed at https://github.com/solitaryzero/NIL_EL

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Fangwei Zhu (10 papers)
  2. Jifan Yu (49 papers)
  3. Hailong Jin (6 papers)
  4. Juanzi Li (144 papers)
  5. Lei Hou (127 papers)
  6. Zhifang Sui (89 papers)
Citations (7)
Github Logo Streamline Icon: https://streamlinehq.com

GitHub