Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Systematic Investigation of KB-Text Embedding Alignment at Scale (2106.01586v1)

Published 3 Jun 2021 in cs.CL and cs.LG

Abstract: Knowledge bases (KBs) and text often contain complementary knowledge: KBs store structured knowledge that can support long range reasoning, while text stores more comprehensive and timely knowledge in an unstructured way. Separately embedding the individual knowledge sources into vector spaces has demonstrated tremendous successes in encoding the respective knowledge, but how to jointly embed and reason with both knowledge sources to fully leverage the complementary information is still largely an open problem. We conduct a large-scale, systematic investigation of aligning KB and text embeddings for joint reasoning. We set up a novel evaluation framework with two evaluation tasks, few-shot link prediction and analogical reasoning, and evaluate an array of KB-text embedding alignment methods. We also demonstrate how such alignment can infuse textual information into KB embeddings for more accurate link prediction on emerging entities and events, using COVID-19 as a case study.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Vardaan Pahuja (14 papers)
  2. Yu Gu (218 papers)
  3. Wenhu Chen (134 papers)
  4. Mehdi Bahrami (7 papers)
  5. Lei Liu (332 papers)
  6. Wei-Peng Chen (6 papers)
  7. Yu Su (138 papers)
Citations (2)