Papers
Topics
Authors
Recent
2000 character limit reached

ReLiK: Retrieve and LinK, Fast and Accurate Entity Linking and Relation Extraction on an Academic Budget

Published 31 Jul 2024 in cs.CL and cs.AI | (2408.00103v3)

Abstract: Entity Linking (EL) and Relation Extraction (RE) are fundamental tasks in Natural Language Processing, serving as critical components in a wide range of applications. In this paper, we propose ReLiK, a Retriever-Reader architecture for both EL and RE, where, given an input text, the Retriever module undertakes the identification of candidate entities or relations that could potentially appear within the text. Subsequently, the Reader module is tasked to discern the pertinent retrieved entities or relations and establish their alignment with the corresponding textual spans. Notably, we put forward an innovative input representation that incorporates the candidate entities or relations alongside the text, making it possible to link entities or extract relations in a single forward pass and to fully leverage pre-trained LLMs contextualization capabilities, in contrast with previous Retriever-Reader-based methods, which require a forward pass for each candidate. Our formulation of EL and RE achieves state-of-the-art performance in both in-domain and out-of-domain benchmarks while using academic budget training and with up to 40x inference speed compared to competitors. Finally, we show how our architecture can be used seamlessly for Information Extraction (cIE), i.e. EL + RE, and setting a new state of the art by employing a shared Reader that simultaneously extracts entities and relations.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (55)
  1. Entity commonsense representation for neural abstractive summarization. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 697–707, New Orleans, Louisiana. Association for Computational Linguistics.
  2. Graph pre-training for AMR parsing and generation. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 6001–6015, Dublin, Ireland. Association for Computational Linguistics.
  3. One spring to rule them both: Symmetric amr semantic parsing and generation without a complex pipeline. Proceedings of the AAAI Conference on Artificial Intelligence, 35(14):12564–12573.
  4. Samuel Broscheit. 2019. Investigating entity knowledge in BERT with simple neural end-to-end entity linking. In Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL), pages 677–685, Hong Kong, China. Association for Computational Linguistics.
  5. Reading Wikipedia to answer open-domain questions. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1870–1879, Vancouver, Canada. Association for Computational Linguistics.
  6. Scalable knowledge graph construction from text collections. In Proceedings of the Second Workshop on Fact Extraction and VERification (FEVER), pages 39–46, Hong Kong, China. Association for Computational Linguistics.
  7. ELECTRA: pre-training text encoders as discriminators rather than generators. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. OpenReview.net.
  8. Highly parallel autoregressive entity linking with discriminative correction. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, Virtual Event / Punta Cana, Dominican Republic, 7-11 November, 2021, pages 7662–7669. Association for Computational Linguistics.
  9. Autoregressive entity retrieval. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. OpenReview.net.
  10. Analysis of named entity recognition and linking for tweets. Inf. Process. Manag., 51(2):32–49.
  11. Faithful to the document or to the world? mitigating hallucinations via entity-linked knowledge in abstractive summarization. In Findings of the Association for Computational Linguistics: EMNLP 2022, pages 1067–1082, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
  12. William Falcon and The PyTorch Lightning team. 2019. PyTorch Lightning.
  13. Exploiting entity linking in queries for entity retrieval. In Proceedings of the 2016 ACM on International Conference on the Theory of Information Retrieval, ICTIR 2016, Newark, DE, USA, September 12- 6, 2016, pages 209–218. ACM.
  14. Debertav3: Improving deberta using electra-style pre-training with gradient-disentangled embedding sharing. In The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023. OpenReview.net.
  15. Kore: Keyphrase overlap relatedness for entity disambiguation. In Proceedings of the 21st ACM International Conference on Information and Knowledge Management, CIKM ’12, page 545–554, New York, NY, USA. Association for Computing Machinery.
  16. Robust disambiguation of named entities in text. In Proceedings of the 2011 Conference on Empirical Methods in Natural Language Processing, pages 782–792, Edinburgh, Scotland, UK. Association for Computational Linguistics.
  17. Pere-Lluís Huguet Cabot and Roberto Navigli. 2021. REBEL: Relation extraction by end-to-end language generation. In Findings of the Association for Computational Linguistics: EMNLP 2021, pages 2370–2381, Punta Cana, Dominican Republic. Association for Computational Linguistics.
  18. A survey on knowledge graphs: Representation, acquisition, and applications. IEEE Trans. Neural Networks Learn. Syst., 33(2):494–514.
  19. GenIE: Generative information extraction. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 4626–4643, Seattle, United States. Association for Computational Linguistics.
  20. Dense passage retrieval for open-domain question answering. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 6769–6781, Online. Association for Computational Linguistics.
  21. End-to-end neural entity linking. In Proceedings of the 22nd Conference on Computational Natural Language Learning, pages 519–529, Brussels, Belgium. Association for Computational Linguistics.
  22. Vision, deduction and alignment: An empirical study on multi-modal knowledge graph alignment. In ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 1–5. IEEE.
  23. RexUIE: A recursive method with explicit schema instructor for universal information extraction. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 15342–15359, Singapore. Association for Computational Linguistics.
  24. On the variance of the adaptive learning rate and beyond. In Proceedings of the Eighth International Conference on Learning Representations (ICLR 2020).
  25. K-bert: Enabling language representation with knowledge graph. Proceedings of the AAAI Conference on Artificial Intelligence, 34(03):2901–2908.
  26. Universal information extraction as unified semantic matching. Proceedings of the AAAI Conference on Artificial Intelligence, 37(11):13318–13326.
  27. Unified structure generation for universal information extraction. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 5755–5772, Dublin, Ireland. Association for Computational Linguistics.
  28. Joint learning of named entity recognition and entity linking. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop, pages 190–196, Florence, Italy. Association for Computational Linguistics.
  29. Entity linking meets word sense disambiguation: a unified approach. Transactions of the Association for Computational Linguistics, 2:231–244.
  30. Open knowledge extraction challenge. In Semantic Web Evaluation Challenges - Second SemWebEval Challenge at ESWC 2015, Portorož, Slovenia, May 31 - June 4, 2015, Revised Selected Papers, volume 548 of Communications in Computer and Information Science, pages 3–15. Springer.
  31. PyTorch: An Imperative Style, High-Performance Deep Learning Library. Curran Associates Inc., Red Hook, NY, USA.
  32. Personalized page rank for named entity disambiguation. In Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 238–243, Denver, Colorado. Association for Computational Linguistics.
  33. KILT: a benchmark for knowledge intensive language tasks. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 2523–2544, Online. Association for Computational Linguistics.
  34. Entity disambiguation with entity definitions. In Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics, pages 1297–1303, Dubrovnik, Croatia. Association for Computational Linguistics.
  35. Modeling relations and their mentions without labeled text. In Machine Learning and Knowledge Discovery in Databases, pages 148–163, Berlin, Heidelberg. Springer Berlin Heidelberg.
  36. N3 - a collection of datasets for named entity recognition and disambiguation in the NLP interchange format. In Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC’14), pages 3529–3533, Reykjavik, Iceland. European Language Resources Association (ELRA).
  37. GERBIL - benchmarking named entity recognition and linking consistently. Semantic Web, 9(5):605–625.
  38. Knowgl: Knowledge generation and linking from text. In Proceedings of the AAAI Conference on Artificial Intelligence.
  39. Dan Roth and Wen-tau Yih. 2004. A linear programming formulation for global inference in natural language tasks. In Proceedings of the Eighth Conference on Computational Natural Language Learning (CoNLL-2004) at HLT-NAACL 2004, pages 1–8, Boston, Massachusetts, USA. Association for Computational Linguistics.
  40. Automatically identifying words that can serve as labels for few-shot text classification. In Proceedings of the 28th International Conference on Computational Linguistics, pages 5569–5578, Barcelona, Spain (Online). International Committee on Computational Linguistics.
  41. Nadine Steinmetz and Harald Sack. 2013. Semantic multimedia information retrieval based on contextual descriptions. In The Semantic Web: Semantics and Big Data, pages 382–396, Berlin, Heidelberg. Springer Berlin Heidelberg.
  42. Joint entity and relation extraction with set prediction networks. IEEE Transactions on Neural Networks and Learning Systems, pages 1–12.
  43. Erik F. Tjong Kim Sang and Fien De Meulder. 2003. Introduction to the CoNLL-2003 shared task: Language-independent named entity recognition. In Proceedings of the Seventh Conference on Natural Language Learning at HLT-NAACL 2003, pages 142–147.
  44. Neural relation extraction for knowledge base enrichment. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 229–240, Florence, Italy. Association for Computational Linguistics.
  45. Rel: An entity linker standing on the shoulders of giants. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR ’20, page 2197–2200, New York, NY, USA. Association for Computing Machinery.
  46. Text embeddings by weakly-supervised contrastive pre-training. arXiv preprint arXiv:2212.03533.
  47. Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 38–45, Online. Association for Computational Linguistics.
  48. Scalable zero-shot entity linking with dense entity retrieval. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 6397–6407, Online. Association for Computational Linguistics.
  49. Word-entity duet representations for document ranking. In Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, Shinjuku, Tokyo, Japan, August 7-11, 2017, pages 763–772. ACM.
  50. LUKE: Deep contextualized entity representations with entity-aware self-attention. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 6442–6454, Online. Association for Computational Linguistics.
  51. Tempel: Linking dynamically evolving and newly emerging entities. In Thirty-sixth Conference on Neural Information Processing Systems Datasets and Benchmarks Track.
  52. A survey for efficient open domain question answering. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 14447–14465, Toronto, Canada. Association for Computational Linguistics.
  53. EntQA: Entity linking as question answering. In International Conference on Learning Representations.
  54. PRGC: Potential relation and global correspondence based joint relational triple extraction. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 6225–6235, Online. Association for Computational Linguistics.
  55. Wenxuan Zhou and Muhao Chen. 2022. An improved baseline for sentence-level relation extraction. In Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pages 161–168, Online only. Association for Computational Linguistics.
Citations (3)

Summary

  • The paper introduces ReLiK, a unified architecture that integrates retrieval and reading to streamline entity linking and relation extraction tasks.
  • It leverages dense passage retrieval and a single forward pass with DeBERTa-v3, achieving EL processing up to 40 times faster than competitors.
  • Empirical results show state-of-the-art performance with improvements up to 8.3 percentage points, making the system accessible for research groups with limited resources.

The paper "ReLiK: Retrieve and Link, Fast and Accurate Entity Linking and Relation Extraction on an Academic Budget" introduces a novel Retriever-Reader architecture named ReLiK. This architecture is designed for efficient and effective Entity Linking (EL) and Relation Extraction (RE), and it integrates these tasks seamlessly within a unified framework. The authors aim to address the limitations of existing EL and RE systems by focusing on three fundamental properties: inference speed, flexibility, and performance. ReLiK achieves state-of-the-art results while being computationally efficient, making it accessible for research groups with limited resources.

Key Innovations and Contributions

ReLiK is composed of two main components: the Retriever and the Reader.

  1. Retriever: This component identifies candidate entities or relations from a given input text. It employs a dense passage retrieval paradigm to encode queries (input text) and passages (textual representations of entities or relations) into dense vector representations. The Retriever ranks the most relevant entities or relations based on these dense representations.
  2. Reader: This component takes the input text along with the retrieved candidate entities or relations and establishes connections between the textual spans and the entities or relations. Importantly, the Reader performs this task in a single forward pass, leveraging modern LLMs such as DeBERTa-v3 to efficiently contextualize the input along with the candidate entities or relations.

Strong Numerical Results

The paper reports strong numerical results across several benchmarks:

  • Entity Linking: ReLiK sets new state-of-the-art results on both in-domain (AIDA) and out-of-domain datasets (MSNBC, Derczynski, KORE50, N3-Reuters-128, N3-RSS-500, OKE-15, OKE-16). Specifically, the large version of ReLiK (ReLiK<sub>L</sub>) achieves the highest performance, significantly outperforming previous state-of-the-art systems by demonstrating an improvement of up to 8.3 percentage points on certain out-of-domain datasets.
  • Relation Extraction: ReLiK shows superior performance on popular RE benchmarks such as NYT and CONLL04, matching or surpassing other state-of-the-art systems. ReLiK is particularly notable for its performance on the REBEL dataset for closed Information Extraction (cIE), achieving high F1 scores for both EL and RE components.

Comparative Analysis and Implications

Compared to existing EL systems, particularly those that rely on sequence-to-sequence models or retriever-reader pairs, ReLiK's unified architecture offers substantial efficiency gains. For example, the system achieves a remarkable advancement in speed, processing EL tasks up to 40 times faster than its closest competitors due to its ability to handle multiple candidate entities in a single forward pass. This efficiency makes ReLiK suitable for real-time and large-scale applications.

Moreover, the flexible design of the Retriever component enhances the model’s capability to handle unseen entities or relations, ensuring robust performance across diverse datasets and scenarios. This flexibility is particularly critical for out-of-domain tasks where traditional methods often fail.

Speculations on Future Developments

Given ReLiK's strong performance and efficiency, future developments could explore its application to broader Information Extraction (IE) tasks beyond EL and RE. Potential areas include automatic text summarization, knowledge base construction, and semantic parsing. Moreover, integrating dynamic entity and relation indices, as well as leveraging continual learning paradigms, could enhance the model's adaptability to evolving knowledge bases.

Furthermore, addressing challenges such as emerging entities and the automated generation of textual representations for entities and relations can reduce dependency on static knowledge bases, fostering improvements in the Retriever component's effectiveness.

Conclusion

ReLiK represents a significant step forward in the field of Information Extraction, offering a fast, flexible, and high-performing solution for Entity Linking and Relation Extraction. The architecture's efficiency allows for cost-effective training and deployment, making state-of-the-art IE accessible to a wider range of research groups and practical applications. By unifying EL and RE within a single framework, ReLiK sets the stage for future innovations that could further enhance the capabilities and applicability of Information Extraction systems.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 3 tweets with 28 likes about this paper.