Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Task-aware Retrieval with Instructions (2211.09260v2)

Published 16 Nov 2022 in cs.CL

Abstract: We study the problem of retrieval with instructions, where users of a retrieval system explicitly describe their intent along with their queries. We aim to develop a general-purpose task-aware retrieval system using multi-task instruction tuning, which can follow human-written instructions to find the best documents for a given query. We introduce the first large-scale collection of approximately 40 retrieval datasets with instructions, BERRI, and present TART, a multi-task retrieval system trained on BERRI with instructions. TART shows strong capabilities to adapt to a new retrieval task via instructions and advances the state of the art on two zero-shot retrieval benchmarks, BEIR and LOTTE, outperforming models up to three times larger. We further introduce a new evaluation setup, X2-Retrieval to better reflect real-world scenarios, where diverse domains and tasks are pooled and a system needs to find documents aligning users' intents. In this setup, TART significantly outperforms competitive baselines, further demonstrating the effectiveness of guiding retrieval with instructions.

Citations (75)

Summary

  • The paper introduces a novel retrieval paradigm using explicit instructions with the BERRI dataset and TART system to enhance zero-shot performance.
  • It compares dense dual-encoder and cross-encoder architectures, demonstrating improved instruction-following and relevance estimation.
  • Empirical evaluations show that instruction tuning, including careful negative sampling, significantly refines task alignment and user intent across diverse domains.

Task-aware Retrieval with Instructions: An Expert Overview

Information retrieval (IR) is a cornerstone of the internet era, enabling users to locate relevant documents from vast collections based on textual queries. However, traditional retrieval systems often exhibit limitations when queries are ambiguous or when implicit user intent is diverse and multifaceted. The paper "Task-aware Retrieval with Instructions" explores a novel approach in IR: retrieval with explicit instructions designed to capture user intent beyond the query itself. This paradigm aims to enhance retrieval precision across a variety of tasks and domains by leveraging the power of instructions.

BERRI and TART: Supporting Retrieval with Instructions

To facilitate research in this domain, the authors introduce BERRI (Bank of Explicit Retrieval Instructions), a comprehensive dataset comprising approximately 40 retrieval tasks enriched with instructions. These instructions encapsulate user intents with insights into the task, domain, and unit of interest. Alongside BERRI, they develop TART (Task-aware ReTriever), a multi-task retrieval system trained on these datasets using instruction tuning.

TART demonstrates its adaptability across new retrieval tasks through instructions, achieving competitive results on zero-shot benchmarks such as BEIR and LOTTE—outperforming models significantly larger in scale. The paper also introduces a new evaluation setup, termed "X", which pools diverse tasks and domains, requiring systems to discern documents according to user intents explicitly communicated via instructions.

Methodology and System Architectures

The authors employ two architectures for TART: a dense dual-encoder, TART-dual, and a cross-encoder, TART-full. TART-dual encodes queries and documents independently, measuring relevance by the similarity between embeddings. In contrast, TART-full combines cross-attention features, performing richer, more expressive interactions between input sequences. By harnessing the pretrained T5 encoder variants for TART-full, the retrieval system benefits from instruction-following capabilities enhanced during model pretraining.

Instruction Tuning and Negative Sampling

A keystone of building resilient retrieval models is the use of varied natural language instructions during training, allowing systems to generalize to unseen tasks. To ensure robustness, the authors crafted hard negatives and instruction-unfollowing negatives, key in compelling TART to focus on task objectives defined via instructions.

Empirical Evaluation and Implications

TART advances state-of-the-art results for zero-shot retrieval, underscoring the efficacy of instructions in refining task alignment and domain understanding without additional data generation requirements. In challenging open-domain settings, TART displays resilience in cross-task retrieval. The paper highlights crucial factors in system development: comprehensive instruction feedback during training, diverse dataset inclusion, substantial model scale, and carefully curated negative samples.

Future Directions and Technical Challenges

The paper posits two future goals: enhancing efficiency in instruction-aware retrievers, especially bi-encoders, and expanding the number of annotated datasets. Further scaling can potentially improve generalization and task awareness in larger machine learning landscapes. Community endeavors to expand dataset collections and instructions are called for, aligning with other instruction-driven model fine-tuning trends.

Overall, “Task-aware Retrieval with Instructions” represents an important evolution in information retrieval methodologies. By introducing explicit user instructions into retrieval processes, it opens avenues not just for improved document alignment but presents opportunities for expanding human-machine interface paradigms in IR.

Youtube Logo Streamline Icon: https://streamlinehq.com