2000 character limit reached
SoftQE: Learned Representations of Queries Expanded by LLMs (2402.12663v1)
Published 20 Feb 2024 in cs.CL, cs.IR, and cs.LG
Abstract: We investigate the integration of LLMs into query encoders to improve dense retrieval without increasing latency and cost, by circumventing the dependency on LLMs at inference time. SoftQE incorporates knowledge from LLMs by mapping embeddings of input queries to those of the LLM-expanded queries. While improvements over various strong baselines on in-domain MS-MARCO metrics are marginal, SoftQE improves performance by 2.83 absolute percentage points on average on five out-of-domain BEIR tasks.
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.