Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

RelBERT: Embedding Relations with Language Models (2310.00299v2)

Published 30 Sep 2023 in cs.CL

Abstract: Many applications need access to background knowledge about how different concepts and entities are related. Although Knowledge Graphs (KG) and LLMs (LLM) can address this need to some extent, KGs are inevitably incomplete and their relational schema is often too coarse-grained, while LLMs are inefficient and difficult to control. As an alternative, we propose to extract relation embeddings from relatively small LLMs. In particular, we show that masked LLMs such as RoBERTa can be straightforwardly fine-tuned for this purpose, using only a small amount of training data. The resulting model, which we call RelBERT, captures relational similarity in a surprisingly fine-grained way, allowing us to set a new state-of-the-art in analogy benchmarks. Crucially, RelBERT is capable of modelling relations that go well beyond what the model has seen during training. For instance, we obtained strong results on relations between named entities with a model that was only trained on lexical relations between concepts, and we observed that RelBERT can recognise morphological analogies despite not being trained on such examples. Overall, we find that RelBERT significantly outperforms strategies based on prompting LLMs that are several orders of magnitude larger, including recent GPT-based models and open source models.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Asahi Ushio (19 papers)
  2. Jose Camacho-Collados (58 papers)
  3. Steven Schockaert (67 papers)
Citations (1)