Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Redefining Information Retrieval of Structured Database via Large Language Models (2405.05508v2)

Published 9 May 2024 in cs.IR and cs.AI

Abstract: Retrieval augmentation is critical when LLMs (LMs) exploit non-parametric knowledge related to the query through external knowledge bases before reasoning. The retrieved information is incorporated into LMs as context alongside the query, enhancing the reliability of responses towards factual questions. Prior researches in retrieval augmentation typically follow a retriever-generator paradigm. In this context, traditional retrievers encounter challenges in precisely and seamlessly extracting query-relevant information from knowledge bases. To address this issue, this paper introduces a novel retrieval augmentation framework called ChatLR that primarily employs the powerful semantic understanding ability of LLMs as retrievers to achieve precise and concise information retrieval. Additionally, we construct an LLM-based search and question answering system tailored for the financial domain by fine-tuning LLM on two tasks including Text2API and API-ID recognition. Experimental results demonstrate the effectiveness of ChatLR in addressing user queries, achieving an overall information retrieval accuracy exceeding 98.8\%.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Mingzhu Wang (3 papers)
  2. Yuzhe Zhang (25 papers)
  3. Qihang Zhao (9 papers)
  4. Hong Zhang (272 papers)
  5. Junyi Yang (30 papers)
X Twitter Logo Streamline Icon: https://streamlinehq.com