Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Towards Disentangling Relevance and Bias in Unbiased Learning to Rank (2212.13937v4)

Published 28 Dec 2022 in cs.IR and cs.AI

Abstract: Unbiased learning to rank (ULTR) studies the problem of mitigating various biases from implicit user feedback data such as clicks, and has been receiving considerable attention recently. A popular ULTR approach for real-world applications uses a two-tower architecture, where click modeling is factorized into a relevance tower with regular input features, and a bias tower with bias-relevant inputs such as the position of a document. A successful factorization will allow the relevance tower to be exempt from biases. In this work, we identify a critical issue that existing ULTR methods ignored - the bias tower can be confounded with the relevance tower via the underlying true relevance. In particular, the positions were determined by the logging policy, i.e., the previous production model, which would possess relevance information. We give both theoretical analysis and empirical results to show the negative effects on relevance tower due to such a correlation. We then propose three methods to mitigate the negative confounding effects by better disentangling relevance and bias. Empirical results on both controlled public datasets and a large-scale industry dataset show the effectiveness of the proposed approaches.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Yunan Zhang (13 papers)
  2. Le Yan (28 papers)
  3. Zhen Qin (105 papers)
  4. Honglei Zhuang (31 papers)
  5. Jiaming Shen (56 papers)
  6. Xuanhui Wang (36 papers)
  7. Michael Bendersky (63 papers)
  8. Marc Najork (27 papers)
Citations (9)

Summary

We haven't generated a summary for this paper yet.