Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Instructed Language Models with Retrievers Are Powerful Entity Linkers (2311.03250v1)

Published 6 Nov 2023 in cs.CL and cs.AI

Abstract: Generative approaches powered by LLMs have demonstrated emergent abilities in tasks that require complex reasoning abilities. Yet the generative nature still makes the generated content suffer from hallucinations, thus unsuitable for entity-centric tasks like entity linking (EL) requiring precise entity predictions over a large knowledge base. We present Instructed Generative Entity Linker (INSGENEL), the first approach that enables casual LLMs to perform entity linking over knowledge bases. Several methods to equip LLMs with EL capability were proposed in this work, including (i) a sequence-to-sequence training EL objective with instruction-tuning, (ii) a novel generative EL framework based on a light-weight potential mention retriever that frees the model from heavy and non-parallelizable decoding, achieving 4$\times$ speedup without compromise on linking metrics. INSGENEL outperforms previous generative alternatives with +6.8 F1 points gain on average, also with a huge advantage in training data efficiency and training compute consumption. In addition, our skillfully engineered in-context learning (ICL) framework for EL still lags behind INSGENEL significantly, reaffirming that the EL task remains a persistent hurdle for general LLMs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Zilin Xiao (9 papers)
  2. Ming Gong (246 papers)
  3. Jie Wu (230 papers)
  4. Xingyao Zhang (17 papers)
  5. Linjun Shou (53 papers)
  6. Jian Pei (104 papers)
  7. Daxin Jiang (138 papers)
Citations (8)

Summary

We haven't generated a summary for this paper yet.