Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Correcting for Selection Bias in Learning-to-rank Systems (2001.11358v2)

Published 29 Jan 2020 in cs.IR

Abstract: Click data collected by modern recommendation systems are an important source of observational data that can be utilized to train learning-to-rank (LTR) systems. However, these data suffer from a number of biases that can result in poor performance for LTR systems. Recent methods for bias correction in such systems mostly focus on position bias, the fact that higher ranked results (e.g., top search engine results) are more likely to be clicked even if they are not the most relevant results given a user's query. Less attention has been paid to correcting for selection bias, which occurs because clicked documents are reflective of what documents have been shown to the user in the first place. Here, we propose new counterfactual approaches which adapt Heckman's two-stage method and accounts for selection and position bias in LTR systems. Our empirical evaluation shows that our proposed methods are much more robust to noise and have better accuracy compared to existing unbiased LTR algorithms, especially when there is moderate to no position bias.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Zohreh Ovaisi (3 papers)
  2. Ragib Ahsan (6 papers)
  3. Yifan Zhang (245 papers)
  4. Kathryn Vasilaky (2 papers)
  5. Elena Zheleva (29 papers)
Citations (95)

Summary

We haven't generated a summary for this paper yet.