All about Sample-Size Calculations for A/B Testing: Novel Extensions & Practical Guide

1Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.

Abstract

While there exists a large amount of literature on the general challenges and best practices for trustworthy online A/B testing, there are limited studies on sample size estimation, which plays a crucial role in trustworthy and efficient A/B testing that ensures the resulting inference has a sufficient power and type I error control. For example, when sample size is under-estimated, the statistical inference, even with the correct analysis methods, will not be able to detect the true significant improvement leading to misinformed and costly decisions. This paper addresses this fundamental gap by developing new sample size calculation methods for correlated data, as well as absolute vs. relative treatment effects, both ubiquitous in online experiments. Additionally, we address a practical question of the minimal observed difference that will be statistically significant and how it relates to average treatment effect and sample size calculation. All proposed methods are accompanied by mathematical proofs, illustrative examples, and simulations. We end by sharing some best practices on various practical topics on sample size calculation and experimental design.

Cite

CITATION STYLE

APA

Zhou, J., Lu, J., & Shallah, A. (2023). All about Sample-Size Calculations for A/B Testing: Novel Extensions & Practical Guide. In International Conference on Information and Knowledge Management, Proceedings (pp. 3574–3583). Association for Computing Machinery. https://doi.org/10.1145/3583780.3614779

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free