When Inverse Propensity Scoring does not Work: Affine Corrections for Unbiased Learning to Rank

58Citations
Citations of this article
31Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Besides position bias, which has been well-studied, trust bias is another type of bias prevalent in user interactions with rankings: users are more likely to click incorrectly w.r.t. their preferences on highly ranked items because they trust the ranking system. While previous work has observed this behavior in users, we prove that existing Counterfactual Learning to Rank (CLTR) methods do not remove this bias, including methods specifically designed to mitigate this type of bias. Moreover, we prove that Inverse Propensity Scoring (IPS) is principally unable to correct for trust bias under non-trivial circumstances. Our main contribution is a new estimator based on affine corrections: it both reweights clicks and penalizes items displayed on ranks with high trust bias. Our estimator is the first estimator that is proven to remove the effect of both trust bias and position bias. Furthermore, we show that our estimator is a generalization of the existing (CLTR) framework: if no trust bias is present, it reduces to the original (IPS) estimator. Our semi-synthetic experiments indicate that by removing the effect of trust bias in addition to position bias, (CLTR) can approximate the optimal ranking system even closer than previously possible.

Cite

CITATION STYLE

APA

Vardasbi, A., Oosterhuis, H., & De Rijke, M. (2020). When Inverse Propensity Scoring does not Work: Affine Corrections for Unbiased Learning to Rank. In International Conference on Information and Knowledge Management, Proceedings (pp. 1475–1484). Association for Computing Machinery. https://doi.org/10.1145/3340531.3412031

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free