LLM-QE: Enhancing Query Expansion with Ranking-Aligned LLMs
The paper presents a novel approach, LLM-QE, for query expansion in information retrieval, leveraging the capabilities of LLMs. This method aims to bridge the semantic gap between query terms and document content, a prevalent challenge in dense retrieval models. Unlike traditional methods, LLM-QE incorporates rank-based and answer-based reward models to align LLMs with the ranking preferences of both retrievers and the LLMs themselves. This mitigation of LLM hallucinations significantly enhances the performance of dense retrieval models, as evidenced by experimental results on the zero-shot dense retrieval model, Contriever. LLM-QE achieved over 8% improvement in retrieval effectiveness, underscoring the benefits of its innovative reward modeling strategy.
The architecture of LLM-QE diverges from standard generative relevance feedback methods, which primarily utilize LLMs for query expansion by directly generating relevant documents or reasoning through steps like Chain-of-Thought (CoT). The authors propose integrating Direct Preference Optimization (DPO) to fine-tune LLMs, thereby crafting more precise and contextually relevant expansions. This process mitigates the risk of irrelevant content introduction, a common shortfall observed when LLMs are directly deployed for query-related content generation.
Another critical component of LLM-QE is its unique reward model. By employing rank-based rewards, the model emulates dense retrieval systems' document ranking preferences. In concert, the answer-based rewards assess the relevance of generated expansions by comparing them to optimized LLM outputs, thus achieving a holistic alignment of expansions with effective document ranking strategies. The results demonstrate LLM-QE's superiority in ensuring query expansions are not only relevant but also concise, avoiding the generation of excessive and unnecessary tokens.
The evaluation conducted on the BEIR dataset reveals significant improvements in both unsupervised and supervised settings, with marked performance boosts noted in complex query-response environments, particularly in enhancing the training of dense retrievers. The robust performance across numerous datasets solidifies LLM-QE as an essential tool for query expansion, especially in scenarios deficient in high-quality training signals.
The implications of this research are multifaceted. Practically, LLM-QE equips information retrieval systems with enhanced tools for addressing semantic discrepancies between queries and large-scale document collections, thereby improving user experience. Theoretically, the framework offers a potent strategy for aligning generative models with ranking paradigms, paving the way for future exploration into more sophisticated integration of LLMs in diverse retrieval contexts. Future research could expand upon this framework, exploring its adaptability to various retrieval models and further refining the interaction between generation and retrieval components in information systems. The open-source availability of all codes ensures that LLM-QE can serve as a foundational model for future developments in query expansion and information retrieval processes, potentially influencing the broader landscape of artificial intelligence applications in data retrieval and interaction.