In Experimentation, we obsess over A/B tests (Control vs. Variant). But the most important test you will ever run is an A/A Test.
An A/A Test splits traffic exactly like an A/B test (50/50), but there is zero difference between the two groups. Everyone sees the Control experience.
Why would you waste traffic on this? Because if your tool reports a "Statistically Significant Winner" in an A/A test, you know your tool is broken.
A/A tests are diagnostic tools. They validate three critical components of your infrastructure:
When you run an A/A test, you are testing the Null Hypothesis ($H_0$) where you know the Null is true.
You run the test. The result shows "No Significant Difference" (p > 0.05). The conversion rates are nearly identical.
Result: PASS ✅You run the test. The tool says "Variant A2 is the Winner!" with 99% confidence.
Result: FAIL ❌ (Bias Detected)You don't need to run an A/A test every week. That is a waste of traffic. You should run them at specific milestones:
A/A tests are also great for calculating the Baseline Variance ($\sigma$). Knowing exactly how much your metric naturally bounces around helps you calculate Sample Size (Module 04) more accurately.
Now that our engine is calibrated, we are ready to run real tests. But running the test is only half the battle. We must ensure the traffic stays balanced. This leads us to the most common error in execution: Sample Ratio Mismatch (SRM).