Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

MILL: Mutual Verification with Large Language Models for Zero-Shot Query Expansion (2310.19056v3)

Published 29 Oct 2023 in cs.IR, cs.AI, and cs.CL

Abstract: Query expansion, pivotal in search engines, enhances the representation of user information needs with additional terms. While existing methods expand queries using retrieved or generated contextual documents, each approach has notable limitations. Retrieval-based methods often fail to accurately capture search intent, particularly with brief or ambiguous queries. Generation-based methods, utilizing LLMs, generally lack corpus-specific knowledge and entail high fine-tuning costs. To address these gaps, we propose a novel zero-shot query expansion framework utilizing LLMs for mutual verification. Specifically, we first design a query-query-document generation method, leveraging LLMs' zero-shot reasoning ability to produce diverse sub-queries and corresponding documents. Then, a mutual verification process synergizes generated and retrieved documents for optimal expansion. Our proposed method is fully zero-shot, and extensive experiments on three public benchmark datasets are conducted to demonstrate its effectiveness over existing methods. Our code is available online at https://github.com/Applied-Machine-Learning-Lab/MILL to ease reproduction.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Pengyue Jia (22 papers)
  2. Yiding Liu (30 papers)
  3. Xiangyu Zhao (192 papers)
  4. Xiaopeng Li (166 papers)
  5. Changying Hao (4 papers)
  6. Shuaiqiang Wang (68 papers)
  7. Dawei Yin (165 papers)
Citations (7)