Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning to Plan for Retrieval-Augmented Large Language Models from Knowledge Graphs (2406.14282v3)

Published 20 Jun 2024 in cs.CL and cs.AI

Abstract: Improving the performance of LLMs in complex question-answering (QA) scenarios has always been a research focal point. Recent studies have attempted to enhance LLMs' performance by combining step-wise planning with external retrieval. While effective for advanced models like GPT-3.5, smaller LLMs face challenges in decomposing complex questions, necessitating supervised fine-tuning. Previous work has relied on manual annotation and knowledge distillation from teacher LLMs, which are time-consuming and not accurate enough. In this paper, we introduce a novel framework for enhancing LLMs' planning capabilities by using planning data derived from knowledge graphs (KGs). LLMs fine-tuned with this data have improved planning capabilities, better equipping them to handle complex QA tasks that involve retrieval. Evaluations on multiple datasets, including our newly proposed benchmark, highlight the effectiveness of our framework and the benefits of KG-derived planning data.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (13)
  1. Junjie Wang (164 papers)
  2. Mingyang Chen (45 papers)
  3. Binbin Hu (42 papers)
  4. Dan Yang (60 papers)
  5. Ziqi Liu (78 papers)
  6. Yue Shen (243 papers)
  7. Peng Wei (112 papers)
  8. Zhiqiang Zhang (129 papers)
  9. Jinjie Gu (50 papers)
  10. Jun Zhou (370 papers)
  11. Jeff Z. Pan (78 papers)
  12. Wen Zhang (170 papers)
  13. Huajun Chen (198 papers)
Citations (6)
X Twitter Logo Streamline Icon: https://streamlinehq.com