how can we actually enable CUPED for our experiments?
The knowledge sources explain what CUPED is and when to use it, but they do not contain explicit step-by-step instructions on how to enable CUPED within the Mixpanel UI for your experiments. [Advanced Statistical Methods]
What the documentation does confirm is that CUPED is one of the advanced statistical methods available in the Experiments report, and that it uses a pre-exposure period of your choosing to adjust for pre-experiment user behavior. [CUPED]
For precise UI instructions on enabling CUPED, I'd recommend:
Checking the full Mixpanel Experiments documentation directly for any UI walkthrough.
Reaching out to your Mixpanel account team, especially since Experiments is an Enterprise add-on.
What the docs do tell us about CUPED:
It works best when you have sufficient pre-experiment data for most users, high-variance metrics, and metrics where past behavior predicts future behavior (e.g., engagement, revenue, retention). [CUPED]
Users without pre-experiment data are assigned a value of zero, so they are still included in the analysis. [CUPED]
It can also help mitigate pre-existing group bias flagged by the Retro-AA health check. [Health Checks]
โน๏ธย Tip: Mention @kapa.ai in the thread for follow-ups or DM the app for 1-to-1 help.
Great question. The knowledge sources don't directly address whether segmenting new vs. existing visitors replaces the need for CUPED, but here's what the documentation does tell us that's relevant:
What CUPED does:
CUPED reduces variance by accounting for how users were already behaving before the experiment. It helps isolate the true treatment effect by controlling for pre-existing behavioral differences between groups โ not by segmenting users, but by adjusting the outcome measurement. [CUPED]
Where your segmentation approach helps:
Monitoring new and existing visitors separately is a sound practice โ it avoids survivorship bias and surfaces different behavioral patterns. The documentation explicitly recommends using segmentation (e.g., new vs. returning users) to understand nuances in experiment results. [Reading Results]
Why CUPED may still be useful:
For your existing users segment specifically, CUPED would be most applicable โ these users have pre-experiment history, which is exactly what CUPED leverages. [CUPED]
For new users, CUPED offers little benefit since they have no prior behavioral data (they'd be assigned a value of zero for the pre-exposure metric). [CUPED]
Bottom line: Segmenting new vs. existing visitors and using CUPED serve different purposes โ they're complementary, not interchangeable. Segmentation helps you analyze subgroups correctly; CUPED helps you detect real effects faster within a group by reducing noise. The knowledge sources don't state that one eliminates the need for the other.
โน๏ธย Tip: Mention @kapa.ai in the thread for follow-ups or DM the app for 1-to-1 help.
An "imbalance alert" in the context of Mixpanel Experiments refers to a Sample Ratio Mismatch (SRM) โ a health check that Mixpanel runs automatically on all experiments.
What triggers an SRM (Imbalance) alert?
You get this alert when the actual distribution of users across your variants doesn't match your configured split in a statistically significant way. For example, if you set up a 50/50 split but end up with 60% in control and 40% in treatment, Mixpanel flags this as a potential problem. [Health Checks]
Mixpanel uses a Chi-squared test to determine whether the observed imbalance is likely due to random chance or signals a real issue. [Health Checks]
Common causes
Bugs in your randomization or bucketing logic
Exposure events not firing consistently across variants
Users being reassigned to different variants mid-experiment
Differences in the start times of each variant
Technical issues like broken redirects [Health Checks; SRM Detection]
What should you do?
The recommended action is to:
Pause the experiment
Identify and fix the root cause of the mismatch
Restart the experiment after fixing it [Advanced Statistical Methods]
๐ก As noted in the previous conversation, enabling CUPED can help mitigate the impact of pre-existing group bias, but it doesn't fix a broken randomization setup โ that needs to be resolved at the source.
โน๏ธย Tip: Mention @kapa.ai in the thread for follow-ups or DM the app for 1-to-1 help.
Even if your observed ratio appears to be 50/50, an SRM alert can still fire if the imbalance is detected at a statistically significant level โ meaning even a small but consistent skew can trigger it.
The knowledge sources don't go into further detail about why an SRM alert might appear despite an apparently balanced split, but based on what's documented, here are the most likely explanations to investigate:
Exposure events not firing consistently across variants โ if one variant's exposure event fires more reliably than the other, the recorded split may look balanced while the actual assignment is not. [Health Checks]
Users being reassigned mid-experiment โ if some users switch variants, this can distort the analysis even if the overall count looks even. [Health Checks]
Bugs in bucketing logic โ the assignment may appear balanced on the surface but have subtle issues affecting specific user segments. [SRM Detection]
The recommended course of action remains the same regardless: pause the experiment, investigate the root cause, and restart after fixing it. [Advanced Statistical Methods]
For a deeper diagnosis, I'd recommend checking whether your exposure event ($experiment_started) is firing at the right moment and consistently for both variants โ not at assignment time, but at actual exposure. [Experiments]
โน๏ธย Tip: Mention @kapa.ai in the thread for follow-ups or DM the app for 1-to-1 help.
