UKD: Debiasing Conversion Rate Estimation via Uncertainty-regularized Knowledge Distillation

18Citations
Citations of this article
32Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In online advertising, conventional post-click conversion rate (CVR) estimation models are trained using clicked samples. However, during online serving the models need to estimate for all impression ads, leading to the sample selection bias (SSB) issue. Intuitively, providing reliable supervision signals for unclicked ads is a feasible way to alleviate the SSB issue. This paper proposes an uncertainty-regularized knowledge distillation (UKD) framework to debias CVR estimation via distilling knowledge from unclicked ads. A teacher model learns click-adaptive representations and produces pseudo-conversion labels on unclicked ads as supervision signals. Then a student model is trained on both clicked and unclicked ads with knowledge distillation, performing uncertainty modeling to alleviate the inherent noise in pseudo-labels. Experiments on billion-scale datasets show that UKD outperforms previous debiasing methods. Online results verify that UKD achieves significant improvements.

Cite

CITATION STYLE

APA

Xu, Z., Wei, P., Zhang, W., Liu, S., Wang, L., & Zheng, B. (2022). UKD: Debiasing Conversion Rate Estimation via Uncertainty-regularized Knowledge Distillation. In WWW 2022 - Proceedings of the ACM Web Conference 2022 (pp. 2078–2087). Association for Computing Machinery, Inc. https://doi.org/10.1145/3485447.3512081

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free