Abstract
The rollout of new versions of a feature in modern applications is a manual multi-stage process, as the feature is released to ever larger groups of users, while its performance is carefully monitored. This kind of A/B testing is ubiquitous, but suboptimal, as the monitoring requires heavy human intervention, is not guaranteed to capture consistent, but short-term fuctuations in performance, and is inefcient, as better versions take a long time to reach the full population. In this work we formulate this question as that of expert learning, and give a new algorithm Follow-The-Best-Interval, FTBI, that works in dynamic, non-stationary environments. Our approach is practical, simple, and efcient, and has rigorous guarantees on its performance. Finally, we perform a thorough evaluation on synthetic and real world datasets and show that our approach outperforms current state-of-the-art methods.
Author supplied keywords
Cite
CITATION STYLE
Medina, A. s.Mu oz, Vassilvitiskii, S., & Yin, D. (2018). Online learning for non-stationary A/B tests. In International Conference on Information and Knowledge Management, Proceedings (pp. 317–326). Association for Computing Machinery. https://doi.org/10.1145/3269206.3271718
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.