Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Towards Better Entity Linking with Multi-View Enhanced Distillation (2305.17371v1)

Published 27 May 2023 in cs.CL

Abstract: Dense retrieval is widely used for entity linking to retrieve entities from large-scale knowledge bases. Mainstream techniques are based on a dual-encoder framework, which encodes mentions and entities independently and calculates their relevances via rough interaction metrics, resulting in difficulty in explicitly modeling multiple mention-relevant parts within entities to match divergent mentions. Aiming at learning entity representations that can match divergent mentions, this paper proposes a Multi-View Enhanced Distillation (MVD) framework, which can effectively transfer knowledge of multiple fine-grained and mention-relevant parts within entities from cross-encoders to dual-encoders. Each entity is split into multiple views to avoid irrelevant information being over-squashed into the mention-relevant view. We further design cross-alignment and self-alignment mechanisms for this framework to facilitate fine-grained knowledge distillation from the teacher model to the student model. Meanwhile, we reserve a global-view that embeds the entity as a whole to prevent dispersal of uniform information. Experiments show our method achieves state-of-the-art performance on several entity linking benchmarks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Yi Liu (543 papers)
  2. Yuan Tian (183 papers)
  3. Jianxun Lian (39 papers)
  4. Xinlong Wang (56 papers)
  5. Yanan Cao (34 papers)
  6. Fang Fang (100 papers)
  7. Wen Zhang (170 papers)
  8. Haizhen Huang (18 papers)
  9. Denvy Deng (9 papers)
  10. Qi Zhang (785 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.