The Critical Challenge of using Large-scale Digital Experiment Platforms for Scientific Discovery

In stock
SKU
18201

 Open access

Publication History

Received: June 25, 2022
Revised: April 10, 2023; Spetember 7, 2023; January 20, 2024
Accepted: March 23, 2024
Published as Forthcoming: August 28, 2024
Published as Articles in Advance: Forthcoming
Published in Issue: Forthcoming

https://doi.org/10.25300/MISQ/2024/18201 


This work is licensed under the terms of a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.

Downloadable File
$0.00
Abstract

Robust digital experimentation platforms have become increasingly pervasive at major technology and e-commerce firms worldwide. They allow product managers to use data driven decision-making through online controlled experiments that estimate the average treatment effect (ATE) relative to a status quo control setting and make associated inferences. As demand for experiments continues to grow, orthogonal test planes (OTPs) have become the industry standard for managing the assignment of users to multiple concurrent experimental treatments in companies using large-scale digital experimentation platforms. In recent years, firms have begun to recognize that test planes might be confounding experimental results, but nevertheless, the practical benefits outweigh the costs. However, the uptick in practitioner-led digital experiments has coincided with an increase in academic-industry research partnerships, where large-scale digital experiments are being used to scientifically answer research questions, validate design choices, and/or to derive computational social science-based empirical insights. In such contexts, confounding and biased estimation may have much more pronounced implications for the validity of scientific findings, contributions to theory, building a cumulative literature, and ultimately practice. The purpose of this issues and opinions (I&O) article is to shed light on OTPs – in our experience, most researchers are unaware how such test planes can lead to incorrect inferences. We use a case study conducted at a major e-commerce company to illustrate the extent to which interactions in concurrent experiments can bias ATEs, often making them appear more positive than they actually are. We discuss implications for research, including the distinction between practical industry experiments and academic research, methodological best practices for mitigating such concerns, and transparency and reproducibility considerations stemming from the complexity and opacity of large-scale experimentation platforms. More broadly, we worry that confounding in scientific research due to reliance on large-scale digital experiments, meant to serve a different purpose, is a microcosm of a larger epistemological confounding regarding what constitutes a contribution to scientific knowledge.

Additional Details
Author Ahmed Abbasi, Sriram Somanchi, and Ken Kelley
Year
Volume
Issue
Keywords large-scale digital experimentation, online controlled experiments, type of research, research approach, machine learning, causal inference
Page Numbers
Copyright © 2024 MISQ. All rights reserved.