UnifieR: A Unified Retriever for Large-Scale Retrieval (2205.11194v2)
Abstract: Large-scale retrieval is to recall relevant documents from a huge collection given a query. It relies on representation learning to embed documents and queries into a common semantic encoding space. According to the encoding space, recent retrieval methods based on pre-trained LLMs (PLM) can be coarsely categorized into either dense-vector or lexicon-based paradigms. These two paradigms unveil the PLMs' representation capability in different granularities, i.e., global sequence-level compression and local word-level contexts, respectively. Inspired by their complementary global-local contextualization and distinct representing views, we propose a new learning framework, UnifieR which unifies dense-vector and lexicon-based retrieval in one model with a dual-representing capability. Experiments on passage retrieval benchmarks verify its effectiveness in both paradigms. A uni-retrieval scheme is further presented with even better retrieval quality. We lastly evaluate the model on BEIR benchmark to verify its transferability.
- Tao Shen (87 papers)
- Xiubo Geng (36 papers)
- Chongyang Tao (61 papers)
- Can Xu (98 papers)
- Guodong Long (115 papers)
- Kai Zhang (542 papers)
- Daxin Jiang (138 papers)