Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Encoding Knowledge Graph Entity Aliases in Attentive Neural Network for Wikidata Entity Linking (1912.06214v3)

Published 12 Dec 2019 in cs.CL

Abstract: The collaborative knowledge graphs such as Wikidata excessively rely on the crowd to author the information. Since the crowd is not bound to a standard protocol for assigning entity titles, the knowledge graph is populated by non-standard, noisy, long or even sometimes awkward titles. The issue of long, implicit, and nonstandard entity representations is a challenge in Entity Linking (EL) approaches for gaining high precision and recall. Underlying KG, in general, is the source of target entities for EL approaches, however, it often contains other relevant information, such as aliases of entities (e.g., Obama and Barack Hussein Obama are aliases for the entity Barack Obama). EL models usually ignore such readily available entity attributes. In this paper, we examine the role of knowledge graph context on an attentive neural network approach for entity linking on Wikidata. Our approach contributes by exploiting the sufficient context from a KG as a source of background knowledge, which is then fed into the neural network. This approach demonstrates merit to address challenges associated with entity titles (multi-word, long, implicit, case-sensitive). Our experimental study shows approx 8% improvements over the baseline approach, and significantly outperform an end to end approach for Wikidata entity linking.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Isaiah Onando Mulang (2 papers)
  2. Kuldeep Singh (50 papers)
  3. Akhilesh Vyas (1 paper)
  4. Saeedeh Shekarpour (20 papers)
  5. Maria Esther Vidal (2 papers)
  6. Jens Lehmann (80 papers)
  7. Soren Auer (5 papers)
Citations (8)