Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3219166.3219204acmconferencesArticle/Chapter ViewAbstractPublication PagesecConference Proceedingsconference-collections
abstract

The A/B Testing Problem

Published: 11 June 2018 Publication History

Abstract

"Randomized experiments are increasingly central to innovation in many fields. In the tech sector, major platforms run thousands of experiments (called A/B tests) each year on tens of millions of users at any given time and use the results to screen most product innovations. In the policy and academic circles, governments, nonprofit organizations, and academics use randomized control trials to evaluate social programs and shape public policy. Experiments are not only prevalent, but also highly heterogeneous in design. Policy makers and tech giants typically focus on a "go big" approach, obtaining large sample sizes for a small number of experiments to ensure they that can detect even small benefits of a policy intervention. In contrast, many start-ups and entrepreneurs take a different "go lean" approach, running many small tests and discarding any innovation without outstanding success. The idea is to quickly and cheaply experiment with many ideas, abandon or pivot from ideas that do not work, and scale up ideas that do work. In this paper, we study when each of these approaches is appropriate. To do so, we propose a new framework for optimal experimentation that we call the A/B testing problem. The frameworks also yields an optimal strategy of what innovations to implement and methods to calculate the value of data and experimentation. The key insight is that the optimal experimentation strategy depends crucially on the tails of the distribution of innovation quality, and whether these tails have "black swan" outliers, of innovations with a very large positive or negative impact. The A/B testing problem is as follows. A firm has a set of potential innovations i=1,-,I to implement. The quality ..._i of innovation i is unknown and comes from a distribution G. Quality is independently distributed across innovations. The firm selects a number of users n_i to allocate to an A/B test evaluating innovation i. This yields a signal with mean equal to the true quality of idea i and variance a^2/n_i. The firm is subject to the constraint that the total number of users assigned to experiments is no greater than the number N of users available for experimentation. After seeing the realization of the signals, the firm selects a subset S of ideas to implement. The firm's objective is to maximize the expected sum of the true quality of the ideas that are implemented.

Supplementary Material

MP4 File (p461.mp4)

References

[1]
Robbins, Herbert, "Some aspects of the sequential design of experiments", in Herbert Robbins Selected Papers (Springer, 1985), pp. 169--177.
[2]
Thompson, William R., "On the likelihood that one unknown probability exceeds another in view of the evidence of two samples", Biometrika 25, 3/4 (1933), pp. 285--294.
[3]
Wald, Abraham, "Foundations of a general theory of sequential decision functions", Econometrica, Journal of the Econometric Society (1947), pp. 279--313.

Cited By

View all
  • (2024)Opportunities for Adaptive Experiments to Enable Continuous Improvement in Computer Science EducationProceedings of the 26th Western Canadian Conference on Computing Education10.1145/3660650.3660659(1-7)Online publication date: 2-May-2024
  • (2022)Comment on: “Confidence Intervals for Nonparametric Empirical Bayes Analysis” by Ignatiadis and WagerJournal of the American Statistical Association10.1080/01621459.2022.2102501117:539(1181-1182)Online publication date: 12-Sep-2022
  • (2021)ROI maximization in stochastic online decision-makingProceedings of the 35th International Conference on Neural Information Processing Systems10.5555/3540261.3540961(9152-9166)Online publication date: 6-Dec-2021
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
EC '18: Proceedings of the 2018 ACM Conference on Economics and Computation
June 2018
713 pages
ISBN:9781450358293
DOI:10.1145/3219166
Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 11 June 2018

Check for updates

Author Tags

  1. a/b testing
  2. innovation screening
  3. optimal experimentation strategy

Qualifiers

  • Abstract

Funding Sources

  • Sloan Research Fellowship

Conference

EC '18
Sponsor:

Acceptance Rates

EC '18 Paper Acceptance Rate 70 of 269 submissions, 26%;
Overall Acceptance Rate 664 of 2,389 submissions, 28%

Upcoming Conference

EC '25
The 25th ACM Conference on Economics and Computation
July 7 - 11, 2025
Stanford , CA , USA

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)32
  • Downloads (Last 6 weeks)5
Reflects downloads up to 13 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2024)Opportunities for Adaptive Experiments to Enable Continuous Improvement in Computer Science EducationProceedings of the 26th Western Canadian Conference on Computing Education10.1145/3660650.3660659(1-7)Online publication date: 2-May-2024
  • (2022)Comment on: “Confidence Intervals for Nonparametric Empirical Bayes Analysis” by Ignatiadis and WagerJournal of the American Statistical Association10.1080/01621459.2022.2102501117:539(1181-1182)Online publication date: 12-Sep-2022
  • (2021)ROI maximization in stochastic online decision-makingProceedings of the 35th International Conference on Neural Information Processing Systems10.5555/3540261.3540961(9152-9166)Online publication date: 6-Dec-2021
  • (2021)On Post-selection Inference in A/B TestingProceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining10.1145/3447548.3467129(2743-2752)Online publication date: 14-Aug-2021
  • (2019)Improving Treatment Effect Estimators Through Experiment SplittingThe World Wide Web Conference10.1145/3308558.3313452(285-295)Online publication date: 13-May-2019
  • (2019)Influence Maximization Based Global Structural Properties: A Multi-Armed Bandit ApproachIEEE Access10.1109/ACCESS.2019.29171237(69707-69747)Online publication date: 2019
  • (undefined)Experimentation, Learning, and Appropriability in Early-Stage VenturesSSRN Electronic Journal10.2139/ssrn.3282261

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media