No results for ""
EXPAND ALL
  • Home
  • API docs

Covariate adjustment and CUPED methodology

Read time: 19 minutes
Last edited: Sep 30, 2024
Experimentation is available for all subscription customers

Experimentation is available to all customers on a Developer, Foundation, or Enterprise plan. If you're on an older Pro or Enterprise plan, Experimentation is available as an add-on. To learn more, read about our pricing. To subscribe to a different plan, contact Sales.

This guide includes advanced concepts

This section includes an explanation of advanced statistical concepts. We provide them for informational purposes, but you do not need to understand these concepts to use CUPED for covariate adjustment.

Overview

This guide gives details on the methodology and usage of the CUPED (Controlled experiments Using Pre-Experiment Data) feature for covariate adjustment on the LaunchDarkly Experimentation platform.

Covariate adjustment refers to the use of variables unaffected by treatment, which are known as covariates, for:

  • Variance reduction: reduces the variance of experiment lift estimates, which increases measurement precision and experiment velocity.
  • Bias removal: removes the conditional bias of experiment lift estimates, which increases measurement accuracy.

In mainstream statistics, covariate adjustment is typically performed using Fisher’s (1932) analysis of covariance (ANCOVA) model. In the context of online experimentation, Deng et al. (2013) introduced CUPED (short for Controlled Experiments Using Pre-Experiment Data), which can be thought of as a special case of ANCOVA with the pre-period version of the modeled outcome as a single covariate.

In this guide, we use the terms covariate adjustment, analysis of covariance (ANCOVA), and CUPED interchangeably.

Context

In a randomized experiment, there are three types of variables defined for each experiment unit, such as "user", in a user-randomized experiment:

  • Treatment: a variable indicating treatment for the unit. For example: 1 if the unit is assigned to the "treatment" variation, and 0 if assigned to the "control" variation.
  • Outcomes: post-treatment variables that we want to measure experiment performance on, such as experiment revenue.
  • Covariates: pre-treatment variables that we use to improve our measurement of the outcomes, typically for segmentation and variance reduction, such as pre-experiment revenue.

Outcomes are post-treatment variables. These are variables potentially affected by the treatment, or measured after the treatment is assigned. An example is revenue measured after the user enters the experiment.

Covariates must be pre-treatment variables, which are variables measured before the treatment is assigned, or variables unaffected by treatment. Examples include revenue measured before a user enters the experiment, which is measured before treatment, and gender, which is unaffected by the treatment.

Method

The goal of covariate adjustment is to improve the measurement of an experiment outcome, such as experiment revenue, through the use of prognostic covariates. Prognostic covariates are covariates predictive of the outcome. Pre-experiment revenue is an example of a prognostic covariate, which is typically predictive of experiment revenue. The ANCOVA model, and CUPED in particular, does this by leveraging the correlation, which is the strength of linear relationship, between an outcome and a set of covariates, with the goal of improving measurement precision and accuracy.

Variance Reduction

We illustrate this with a simple example of an outcome , such as experiment revenue, and a covariate , such as pre-experiment revenue.

In this example, there is a strong linear relationship between them for both treatment and control variations, shown in the scatter plot on the left below:

Left: A scatter plot of outcome versus covariate with sample mean prediction lines. Right: A density plot of errors relative to control mean, showing large variance.
Left: A scatter plot of outcome versus covariate with sample mean prediction lines. Right: A density plot of errors relative to control mean, showing large variance.

Predicting the observations in the treatment and control variations with, respectively, the sample means and results in a large variance for the errors, as illustrated in the plot on the right above.

However, we can leverage the linear relationship between and by predicting the observations in the treatment and control variations with, respectively, the regression predictions and , as shown in the scatter plot on the left below:

Left: A scatter plot of outcome versus covariate with regression prediction lines. Right: A density plot of errors relative to the control regression prediction, showing smaller variance.
Left: A scatter plot of outcome versus covariate with regression prediction lines. Right: A density plot of errors relative to the control regression prediction, showing smaller variance.

This results in smaller variance for the errors, as shown in the density plot on the right above. The above two scatter plots were inspired by those shown in Huitema (2011).

The correlation, which is the strength of the linear relationship between the outcome and the covariate , determines how much the error variance is reduced. The larger the correlation, the larger the variance reduction.

Specifically, if we denote the original error variance estimates for the two variations by, respectively, and , and the new error variance estimates using CUPED by, respectively, and , and the outcome-covariate correlations by, respectively, and , then the following holds approximately:

The proportional reduction in error variance is approximately the square of the correlations:

If the correlations in both variations are , the error variance will be reduced by , and if they are , the error variance will be reduced by . The proportional reduction in the error variance translates to about the same proportion reduction in the variance of the experiment lift estimate, which translates to the same proportional reduction in experiment duration on average. Therefore, when the correlations are , the experiment duration will be reduced by as much as on average, and when they are , the experiment duration will be reduced by as much as on average. In other words, this can cut experiment duration nearly in half.

Bias Removal

In addition to reducing the variance of lift estimates, CUPED applies an adjustment to the sample means and to produce the following covariate-adjusted means:

Where and denote the covariate means for, respectively, the treatment and control variations, and denote the covariate mean over all experiment variations. Although the unadjusted means and are unbiased estimators of the variation averages over many realizations of the experiment, for a specific experiment there could be some conditional bias. Conditional bias may occur due to the random imbalances between the treatment and control variation covariate means and . As long as the linear regression model is correct, the adjustments and control for these imbalances and remove the conditional bias.

Implementation

In this section we discuss the scope and model for the CUPED implementation in the LaunchDarkly Experimentation product.

Scope

CUPED is available for:

  • Unsliced results: experiment results that have not been sliced by attribute. CUPED is not applied to sliced results.
  • Average metrics: metrics using the "average" analysis method, regardless of their scale, including conversion metrics and numeric continuous metrics. This excludes metrics using a percentile analysis method.
  • Experiments that have been running for at least 90 minutes: CUPED is not available for experiments receiving events for less than an hour and a half, due to the longer time it takes to compute the covariate-adjusted means and their standard errors (SEs) in the data pipeline. Results before the first hour and a half are not covariate-adjusted, even if the metrics have been enabled for covariate adjustment.

Model

The covariate adjustment model implemented is characterized by the following two features:

  • Feature 1—Most General Model: LaunchDarkly uses the most general ANCOVA model, which allows for unequal covariate slopes and unequal error variances by experiment group. Using the convention of Yang and Tsiatis (2001), we refer to this model as the ANCOVA 3 model. To learn more, read Covariate adjustment.
  • Feature 2—Single Pre-period Covariate: LaunchDarkly restricts the model to use only one covariate. We used the pre-period version of the modeled outcome, which is the covariate proposed by Deng et al. (2013) for the CUPED model and by Soriano (2019) for the PrePost model.

Besides giving us the most general model, another advantage of Feature 1 is that we can implement the ANCOVA model by fitting separate linear regression models by variation. This means we fit one for each experiment variation, which simplifies implementation. One advantage of Feature 2 is that we can fit the linear regression models using simple analytical formulas without needing to use specialized statistical software for linear regression. Combining Features 1 and 2 yields a very simple SQL implementation that you can apply to big data with computational efficiency.

Some may express concern about our using only one covariate in the model when we could potentially include more. In practice, using only the single pre-period covariate is advantageous from both the data collection and model fit points of view:

  • Data collection: it simplifies the data collection process because we do not need to gather more covariates. Instead, we only need to collect data for the current outcome in the pre-period to obtain the pre-period covariate.
  • Model fit: in practice, the pre-period covariate used is typically the covariate with the largest correlation with the experiment-period outcome. When such a highly correlated covariate is included in the model, including additional covariates typically does not improve the overall fit. In other words, the R-squared of the linear regression model would not increase by much.

The pre-period covariate is measured over a seven-day lookback window before the start of the experiment. Precedent for using only seven days is established by the implementation of the PrePost model for covariate adjustment for YouTube experiments, mentioned in Soriano (2019), which is the basis for our implementation.

There is also a tradeoff between using shorter versus longer windows in terms of relevance versus sufficiency. Shorter windows may have more relevance due to the recency of the information measured, but may not have captured all the information to optimize the outcome-covariate correlation. Longer windows capture more information, but risk including irrelevant information from older events, which may decrease the outcome-covariate correlation.

User interface (UI)

The LaunchDarkly experiment results UI displays CUPED information in the following ways:

  • Scope: the UI indicates whether CUPED for covariate adjustment is enabled for the experiment and applied to the experiment metrics.
  • Variance Reduction: the UI shows the proportional reduction in variance for each covariate-adjusted metric.

Scope

The Launchdarkly UI indicates whether CUPED is disabled and enabled for your experiment and whether it has been applied to your metrics, in the following four scenarios:

  • CUPED is disabled: CUPED is disabled for your experiment because all your metrics are percentile metrics.
  • CUPED is enabled. Results have not been adjusted: none of the metrics are covariate-adjusted yet, because the experiment has been receiving results for less than 90 minutes.
  • CUPED is enabled. Covariate-adjusted results displayed for N of N metrics: some of the metrics are covariate-adjusted.
  • CUPED is enabled. All results have been covariate-adjusted: all of the metrics are covariate-adjusted.

If the experiment results are unsliced, the CUPED indicator appears at the top of the results:

The CUPED indicator for unsliced experiment results.
The CUPED indicator for unsliced experiment results.

When the results are sliced, the "CUPED" chip appears next to the unsliced "All Attributes" heading.

The "CUPED not applied" chip appears next to the heading for any sliced results, such as in the "planType: group" slice:

The CUPED indicator for sliced experiment results.
The CUPED indicator for sliced experiment results.

Variance reduction

When viewing experiment results, hovering over the "CUPED" chip displays a message on the percentage of reduction in the variance of the relative difference:

A CUPED message displaying the percentage that the relative difference was reduced by.
A CUPED message displaying the percentage that the relative difference was reduced by.

When there are three or more variations, LaunchDarkly displays the maximum variance reduction among the treatment versus control variation comparisons. The text displays details relevant to your results. For example: "CUPED reduced variance of relative differences by at most 23%."

Advanced topics

For those interested, we will cover some advanced topics in the following sections.

Covariate adjustment

For a two-variation experiment, you can formulate the ANCOVA 3 model implemented at LaunchDarkly as a single model. For example:

where if unit is in the treatment variation and if unit is in the control variation.

The original ANCOVA model introduced by Fisher (1932) makes the following assumptions:

  • Assumption 1—Equal Slopes: equal covariate slope for all experiment variations, that is, in the example above.
  • Assumption 2—Equal Variances: equal error variance for all experiment variations, that is, in the example above.

Yang and Tsiatis (2001) referred to this original model as the ANCOVA 1 model. If we remove Assumption 1 to allow for unequal covariate slopes, that is, allowing for , then we have what Yang and Tsiatis (2001) calls the ANCOVA 2 model, also known as Lin’s (2013) model or the ANHECOVA (ANalysis of HEterogeneous COVAriance) model of Ye et al. (2021).

However, in practice it can be convenient to relax Assumption 2 in addition to Assumption 1, which allows for unequal error variances, that is, . This gives us what we call the ANCOVA 3 model.

This can be implemented in two ways:

  • Single model: a single generalized least squares (GLS) model, which allows for error variances that vary by experiment group. This can be fitted using, for example, the nlme::gls function in R.
  • Separate models: an equivalent, but simpler, way to implement ANCOVA 3 is to fit one separate regression model for each experiment variation.

Fitting separate models has the advantage of fitting very simple regression models when there is only one covariate. This makes for a simple SQL implementation without leveraging additional software, which improves computational efficiency, especially on big data. We give an example of a simple SQL implementation of the ANCOVA 3 model in the section SQL Implementation.

Causal inference

In a comparative study, whether a randomized experiment or an observational study, the goal is to perform causal inference, which includes estimating the causal effect of a treatment, for example, the causal effect of a new product feature on revenue.

Under the Neyman-Rubin potential outcomes framework for causal inference, we begin with individual potential outcomes (IPOs) and for, respectively, receiving the treatment and not receiving the treatment, for each individual . The individual treatment effect (ITE) for individual is given by:

One estimand for the causal effect of treatment is the average treatment effect (ATE), which is the average of the ITEs:

This is the difference between the average potential outcomes (APOs) and of receiving and not receiving the treatment, respectively. An alternate causal estimand is the relative average treatment effect (RATE):

In the LaunchDarkly Experimentation product, we estimate the APO for each experiment variation for every combination of analysis time, experiment iteration, metric, and dimension slice. We then perform causal inference based on estimating the RATE for each treatment variation versus control.

Covariate-adjusted means

To perform causal inference, we first estimate the IPOs by their respective linear regression predictions for the treatment and control variations using the ANCOVA 3 model described earlier:

The APOs are estimated by averaging the IPOs over all available units. In this case, the units are in both the treatment and control variations:

where denotes the average of the covariate over all units in both variations. Because the linear regression models have only one predictor, the estimated regression intercepts are given by:

Therefore, the estimated APOs are given by:

We refer to and as covariate-adjusted means. They are the unadjusted sample means and , minus the adjustments and . This removes conditional bias due to the randomized imbalances between the covariate means and for both the treatment and control variations, respectively.

You can compute the estimated regression slopes with the following formulas:

where:

  • and are the sample standard deviation (SD) for the outcome in the treatment and control variations, respectively
  • and are the sample SD for the covariate in the treatment and control variations, respectively, and
  • and are the outcome-covariate correlation in the treatment and control variations, respectively.

We can show that the estimated SEs for the covariate-adjusted means for both the treatment and control variations are:

where and are the sample sizes for the treatment and control variations, respectively.

When the sample sizes and are large and the imbalances and are negligible, the above SEs reduce to the following:

Therefore, the proportional variance reduction for each is approximately equal to the squared correlation for the variation, as we showed earlier:

Frequentist and Bayesian approaches

For frequentist estimates, the estimates of the APOs are the above covariate-adjusted means and . However, the LaunchDarkly Experimentation product implements a Bayesian model where the APO estimates are regularized using empirical Bayes priors. To learn more, read Experimentation Statistical Methodology.

The Bayesian results without covariate adjustment through CUPED will continue to use the normal-normal model for custom conversion count and custom numeric continuous metrics and the beta-binomial model for custom conversion binary, clicked or tapped, and page viewed metrics. However, the Bayesian results with covariate adjustment through CUPED will use the normal-normal model for all metrics using the "average" analysis method, including custom conversion binary metrics. Under this model, we assume the following prior distribution for the parameter estimated in variation :

For details on the prior mean and , read Experimentation Statistical Methodology.

We are given a frequentist estimate and its estimated standard error . For the non-CUPED results, the estimate is the sample mean. For CUPED results, the estimate in the covariate-adjusted mean , with details provided in the previous section.

We define precision as the inverse of the variance, which is equivalent to the inverse of the squared standard error. Therefore, the estimated precisions of the prior distribution and the frequentist estimate are, respectively:

Define the following precision sum and weight:

Then the posterior distribution of the estimated parameter is given by:

where the posterior mean is given by the precision-weighted average of the frequentist estimate and the prior mean , and the posterior variance is the inverse of the sum of the frequentist estimate precision and the prior precision .

SQL implementation

Here is an example SQL implementation of the ANCOVA 3 model for covariate adjustment to demonstrate its simplicity.

Assume that we have fields y and x in a table named UnitTable, which is aggregated by experiment units, with fields for analysis time, experiment, metric, segment, and variation. The following simple query produces non-CUPED and CUPED estimates with corresponding SEs aggregated by combinations of analysis time, experiment, metric, segment, and variation:

WITH BasicStats AS (
SELECT
analysis_time,
experiment,
metric,
segment,
arm,
COUNT(*) AS n,
AVG(y) AS ybar,
AVG(x) AS xbar,
AVG(x) OVER (PARTITION BY analysis_time, experiment, metric, segment)
AS xbar_all,
STDEV_SAMP(y) AS s_y,
STDEV_SAMP(x) AS s_x,
CORR(x, y) AS r
FROM UnitTable
GROUP BY 1, 2, 3, 4, 5
)
SELECT
analysis_time,
experiment,
metric,
segment,
arm,
'unadjusted' AS method,
n AS exp_unit_count,
ybar AS estimate,
s_y / SQRT(n) AS estimate_std_error
FROM BasicStats
UNION ALL
SELECT
analysis_time,
experiment,
metric,
segment,
arm,
'covariate_adjusted' AS method,
n AS exp_unit_count,
ybar - (r * s_y / s_x) * (xbar - xbar_all) AS estimate,
s_y * SQRT(1 - SQUARE(r)) * SQRT((n - 1) / (n - 2)) *
SQRT(1 / n + SQUARE(xbar - xbar_all) / (SQUARE(s_x) * (n - 1)))
AS estimate_std_error
FROM BasicStats

The BasicStats common table expression (CTE) produces the following aggregated statistics needed to compute the unadjusted and covariate-adjusted means for each combination of analysis time, experiment, metric, segment, and variation:

  • Sample means: the sample means and for the outcome and the covariate, respectively.
  • Sample standard deviations: the sample standard deviations and for the outcome and the covariate, respectively.
  • Sample correlation: the sample correlation between the outcome and the covariate.

The outer query takes the aggregated statistics from the BasicStats CTE to compute the unadjusted and covariate-adjusted means and their SEs using the formulas we derived in the "Covariate-adjusted means" section.

References

Deng, Alex, Ya Xu, Ron Kohavi, and Toby Walker (2013). "Improving the Sensitivity of Online Controlled Experiments by Utilizing Pre-Experiment Data." WSDM’13, Rome, Italy.

Fisher, Ronald A. (1932). Statistical Methods for Research Workers. Oliver and Boyd. Edinburgh, 4th ed.

Huitema, Bradley (2011). Analysis of Covariance and Alternatives: Statistical Methods for Experiments, Quasi-Experiments, and Single-Case Studies, 2nd ed. Wiley.

Lin, Winston (2013). "Agnostic Notes on Regression Adjustments to Experimental Data: Reexamining Freedman’s Critique." Annals of Applied Statistics, 7(1): 295-318.

Soriano, Jacopo (2019). “Percent Change Estimation in Large Scale Online Experiments.” https://arxiv.org/pdf/1711.00562.pdf.

Yang, Li and Anastasios A. Tsiatis. (2001). "Efficiency Study of Estimators for a Treatment Effect in a Pretest-posttest Trial." American Statistician, 55: 314-321.

Ye, Ting, Jun Shao, Yanyao Yi, and Qingyuan Zhao (2023). "Toward Better Practice of Covariate Adjustment in Analyzing Randomized Clinical Trials." Journal of the American Statistical Association, 118(544): 2370-2382.

Want to know more? Start a trial.

Your 14-day trial begins as soon as you sign up. Get started in minutes using the in-app Quickstart. You'll discover how easy it is to release, monitor, and optimize your software.

Want to try it out? Start a trial.