Language Models as Knowledge Embeddings (2206.12617v3)
Abstract: Knowledge embeddings (KE) represent a knowledge graph (KG) by embedding entities and relations into continuous vector spaces. Existing methods are mainly structure-based or description-based. Structure-based methods learn representations that preserve the inherent structure of KGs. They cannot well represent abundant long-tail entities in real-world KGs with limited structural information. Description-based methods leverage textual information and LLMs. Prior approaches in this direction barely outperform structure-based ones, and suffer from problems like expensive negative sampling and restrictive description demand. In this paper, we propose LMKE, which adopts LLMs to derive Knowledge Embeddings, aiming at both enriching representations of long-tail entities and solving problems of prior description-based methods. We formulate description-based KE learning with a contrastive learning framework to improve efficiency in training and evaluation. Experimental results show that LMKE achieves state-of-the-art performance on KE benchmarks of link prediction and triple classification, especially for long-tail entities.
- Tucker: Tensor factorization for knowledge graph completion. arXiv preprint arXiv:1901.09590, 2019.
- Translating embeddings for modeling multi-relational data. In Proc. of NeurIPS, 2013.
- Mem-kgc: Masked entity model for knowledge graph completion with pre-trained language model. IEEE Access, 2021.
- Convolutional 2d knowledge graph embeddings. In Proc. of AAAI, 2018.
- Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
- Are pretrained language models symbolic reasoners over knowledge? In Proc. of CoNLL, 2020.
- Learning entity and relation embeddings for knowledge graph completion. In Proc. of AAAI, 2015.
- George A Miller. Wordnet: a lexical database for english. Communications of the ACM, 1995.
- A three-way model for collective learning on multi-relational data. In Icml, 2011.
- Semi-supervised entity alignment via knowledge graph embedding with awareness of degree difference. In Proc. of WWW, 2019.
- Language models as knowledge bases? arXiv preprint arXiv:1909.01066, 2019.
- Reasoning with neural tensor networks for knowledge base completion. In Advances in neural information processing systems, 2013.
- Rotate: Knowledge graph embedding by relational rotation in complex space. arXiv preprint arXiv:1902.10197, 2019.
- Kristina Toutanova. Observed versus latent features for knowledge base and text inference. ACL-IJCNLP 2015, page 57, 2015.
- Complex embeddings for simple link prediction. In Proc. of ICML, 2016.
- Attention is all you need. In Advances in neural information processing systems, 2017.
- Knowledge graph embedding by translating on hyperplanes. In Proc. of AAAI, 2014.
- Knowledge graph embedding: A survey of approaches and applications. IEEE Transactions on Knowledge and Data Engineering, 2017.
- Coke: Contextualized knowledge graph embedding. arXiv preprint arXiv:1911.02168, 2019.
- Kepler: A unified model for knowledge embedding and pre-trained language representation. arXiv preprint arXiv:1911.06136, 2019.
- Structure-augmented text representation learning for efficient knowledge graph completion. arXiv preprint arXiv:2004.14781, 2020.
- Representation learning of knowledge graphs with entity descriptions. In Proc. of AAAI, 2016.
- Embedding entities and relations for learning and inference in knowledge bases. arXiv preprint arXiv:1412.6575, 2014.
- Kg-bert: Bert for knowledge graph completion. arXiv preprint arXiv:1909.03193, 2019.
- Triple classification for scholarly knowledge graph completion. arXiv preprint arXiv:2111.11845, 2021.
- Learning hierarchy-aware knowledge graph embeddings for link prediction. In Proc. of AAAI, 2020.
- Pretrain-kge: Learning knowledge representation from pretrained language models. In Proc. of EMNLP, 2020.
- Xintao Wang (132 papers)
- Qianyu He (26 papers)
- Jiaqing Liang (62 papers)
- Yanghua Xiao (151 papers)