Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

ChatKBQA: A Generate-then-Retrieve Framework for Knowledge Base Question Answering with Fine-tuned Large Language Models (2310.08975v3)

Published 13 Oct 2023 in cs.CL and cs.AI

Abstract: Knowledge Base Question Answering (KBQA) aims to answer natural language questions over large-scale knowledge bases (KBs), which can be summarized into two crucial steps: knowledge retrieval and semantic parsing. However, three core challenges remain: inefficient knowledge retrieval, mistakes of retrieval adversely impacting semantic parsing, and the complexity of previous KBQA methods. To tackle these challenges, we introduce ChatKBQA, a novel and simple generate-then-retrieve KBQA framework, which proposes first generating the logical form with fine-tuned LLMs, then retrieving and replacing entities and relations with an unsupervised retrieval method, to improve both generation and retrieval more directly. Experimental results show that ChatKBQA achieves new state-of-the-art performance on standard KBQA datasets, WebQSP, and CWQ. This work can also be regarded as a new paradigm for combining LLMs with knowledge graphs (KGs) for interpretable and knowledge-required question answering. Our code is publicly available.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (12)
  1. Haoran Luo (31 papers)
  2. Haihong E (13 papers)
  3. Zichen Tang (14 papers)
  4. Shiyao Peng (3 papers)
  5. Yikai Guo (9 papers)
  6. Wentai Zhang (8 papers)
  7. Chenghao Ma (3 papers)
  8. Guanting Dong (46 papers)
  9. Meina Song (14 papers)
  10. Wei Lin (207 papers)
  11. Yifan Zhu (84 papers)
  12. Luu Anh Tuan (55 papers)
Citations (25)