Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Language Models as Knowledge Embeddings (2206.12617v3)

Published 25 Jun 2022 in cs.CL, cs.AI, and cs.LG

Abstract: Knowledge embeddings (KE) represent a knowledge graph (KG) by embedding entities and relations into continuous vector spaces. Existing methods are mainly structure-based or description-based. Structure-based methods learn representations that preserve the inherent structure of KGs. They cannot well represent abundant long-tail entities in real-world KGs with limited structural information. Description-based methods leverage textual information and LLMs. Prior approaches in this direction barely outperform structure-based ones, and suffer from problems like expensive negative sampling and restrictive description demand. In this paper, we propose LMKE, which adopts LLMs to derive Knowledge Embeddings, aiming at both enriching representations of long-tail entities and solving problems of prior description-based methods. We formulate description-based KE learning with a contrastive learning framework to improve efficiency in training and evaluation. Experimental results show that LMKE achieves state-of-the-art performance on KE benchmarks of link prediction and triple classification, especially for long-tail entities.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (27)
  1. Tucker: Tensor factorization for knowledge graph completion. arXiv preprint arXiv:1901.09590, 2019.
  2. Translating embeddings for modeling multi-relational data. In Proc. of NeurIPS, 2013.
  3. Mem-kgc: Masked entity model for knowledge graph completion with pre-trained language model. IEEE Access, 2021.
  4. Convolutional 2d knowledge graph embeddings. In Proc. of AAAI, 2018.
  5. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
  6. Are pretrained language models symbolic reasoners over knowledge? In Proc. of CoNLL, 2020.
  7. Learning entity and relation embeddings for knowledge graph completion. In Proc. of AAAI, 2015.
  8. George A Miller. Wordnet: a lexical database for english. Communications of the ACM, 1995.
  9. A three-way model for collective learning on multi-relational data. In Icml, 2011.
  10. Semi-supervised entity alignment via knowledge graph embedding with awareness of degree difference. In Proc. of WWW, 2019.
  11. Language models as knowledge bases? arXiv preprint arXiv:1909.01066, 2019.
  12. Reasoning with neural tensor networks for knowledge base completion. In Advances in neural information processing systems, 2013.
  13. Rotate: Knowledge graph embedding by relational rotation in complex space. arXiv preprint arXiv:1902.10197, 2019.
  14. Kristina Toutanova. Observed versus latent features for knowledge base and text inference. ACL-IJCNLP 2015, page 57, 2015.
  15. Complex embeddings for simple link prediction. In Proc. of ICML, 2016.
  16. Attention is all you need. In Advances in neural information processing systems, 2017.
  17. Knowledge graph embedding by translating on hyperplanes. In Proc. of AAAI, 2014.
  18. Knowledge graph embedding: A survey of approaches and applications. IEEE Transactions on Knowledge and Data Engineering, 2017.
  19. Coke: Contextualized knowledge graph embedding. arXiv preprint arXiv:1911.02168, 2019.
  20. Kepler: A unified model for knowledge embedding and pre-trained language representation. arXiv preprint arXiv:1911.06136, 2019.
  21. Structure-augmented text representation learning for efficient knowledge graph completion. arXiv preprint arXiv:2004.14781, 2020.
  22. Representation learning of knowledge graphs with entity descriptions. In Proc. of AAAI, 2016.
  23. Embedding entities and relations for learning and inference in knowledge bases. arXiv preprint arXiv:1412.6575, 2014.
  24. Kg-bert: Bert for knowledge graph completion. arXiv preprint arXiv:1909.03193, 2019.
  25. Triple classification for scholarly knowledge graph completion. arXiv preprint arXiv:2111.11845, 2021.
  26. Learning hierarchy-aware knowledge graph embeddings for link prediction. In Proc. of AAAI, 2020.
  27. Pretrain-kge: Learning knowledge representation from pretrained language models. In Proc. of EMNLP, 2020.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Xintao Wang (132 papers)
  2. Qianyu He (26 papers)
  3. Jiaqing Liang (62 papers)
  4. Yanghua Xiao (151 papers)
Citations (35)