A/B Testing: Why Most Tests Fail Before They Start
The biggest mistake in A/B testing is not choosing the wrong variant — it is running tests without enough data to prove anything. Here is how to plan a test that actually delivers results.
A/B testing sounds simple. Show version A to half your audience, version B to the other half, and pick the winner. In theory, it is the most rational way to make marketing decisions.
In practice, most A/B tests are worthless. Not because the idea is wrong, but because the execution fails at the most fundamental level: there is not enough data to draw any real conclusions.
We see this pattern constantly with our clients. Someone runs a test for a few days, sees that one variant has a slightly higher click rate, declares it the winner, and rolls it out. The problem? The difference was pure noise — statistical randomness that would disappear with a larger sample.
Step Zero: Decide What You Are Actually Measuring
Before you change a single headline or button colour, you need to answer one question: what is your KPI?
This sounds obvious, but it is where most tests go wrong. Different KPIs require dramatically different amounts of data, and picking the wrong one can make your test impossible from the start.
Common A/B testing KPIs:
- Click-through rate (CTR) — How many people click your ad, email, or CTA. High-volume metric, relatively fast to test.
- Cost per click (CPC) — How much you pay per click. Useful for comparing ad efficiency, but influenced by auction dynamics.
- CPM (cost per thousand impressions) — How much it costs to reach 1,000 people. A media buying metric, not a performance metric.
- Conversion rate — How many visitors complete the desired action (purchase, signup, form submission). The gold standard, but requires the most data.
- Cost per conversion (CPA) — How much you pay per conversion. Combines conversion rate with media cost.
- Revenue per visitor — Total revenue divided by total visitors. Captures both conversion rate and order value.
Here is the critical insight: the rarer the event you are measuring, the more data you need.
If your landing page converts at 10%, you might need 400 visitors per variant to detect a meaningful difference. If it converts at 1%, you might need 15,000 visitors per variant. And if you are measuring purchases at 0.5%, you could need 30,000 or more.
This is why choosing your KPI first is not a formality — it determines whether your test is even feasible.
The Statistics: Why Small Samples Lie
Suppose you test two ad headlines. After 200 clicks, Headline A has a 3.5% conversion rate and Headline B has 2.8%. Headline A is better, right?
Not necessarily. With 200 clicks per variant, that difference is well within the range of random chance. The p-value — the probability that this difference occurred by accident — would be around 0.35. That is nowhere near the 0.05 threshold needed for statistical significance.
For context, a p-value of 0.35 means there is a 35% chance you are looking at pure noise. Would you bet your marketing budget on a coin flip that is only slightly tilted?
The commonly used significance levels in A/B testing are:
- p < 0.05 (95% confidence) — the standard threshold. You can be 95% sure the difference is real.
- p < 0.01 (99% confidence) — for high-stakes decisions where being wrong is expensive.
Most online A/B testing calculators default to 95% confidence, and that is a reasonable starting point. But confidence means nothing if your sample is too small to reach it.
How Many Visitors Do You Actually Need?
The sample size depends on three things:
- Your baseline conversion rate — what the page or ad currently converts at.
- The minimum detectable effect (MDE) — the smallest improvement you care about.
- Your confidence level — typically 95%.
Here are realistic examples:
- Baseline 10%, MDE 2 percentage points (10% → 12%) — you need roughly 3,700 visitors per variant (7,400 total).
- Baseline 5%, MDE 1 percentage point (5% → 6%) — you need roughly 7,500 visitors per variant (15,000 total).
- Baseline 2%, MDE 0.5 percentage points (2% → 2.5%) — you need roughly 14,500 visitors per variant (29,000 total).
- Baseline 1%, MDE 0.3 percentage points (1% → 1.3%) — you need roughly 35,000 visitors per variant (70,000 total).
These numbers are not arbitrary — they come from statistical power analysis. If you run a test with fewer visitors, you are essentially guessing.
The Budget Question: Can You Afford This Test?
Now comes the part most people skip. You know your KPI. You know how many visitors you need. The next question is: can your budget deliver that traffic within a reasonable timeframe?
Work backwards from the numbers:
- Calculate your required sample size based on your baseline rate and MDE.
- Estimate your cost per visitor. If you are running paid ads, this is roughly your CPC. If you are testing on organic traffic, calculate your daily visitor count.
- Multiply sample size by cost per visitor. This is your test budget.
- Divide by your daily budget to get duration. If the test would take 6 months, it is not a viable test.
A practical example:
- You want to test two landing pages. Current conversion rate: 3%. You want to detect a 1 percentage point improvement.
- Required sample: approximately 6,000 visitors per variant = 12,000 total.
- Your CPC is 1.50 euros. Test cost: 18,000 euros.
- Your daily ad budget is 200 euros per day. Duration: 90 days.
Is 18,000 euros and 3 months worth detecting a 1 percentage point lift? Maybe. Maybe not. But now you are making an informed decision rather than running a test that was doomed from the start.
If the budget is too high or the duration too long, you have options:
- Increase the MDE. Test for bigger changes instead of marginal improvements. A complete headline rewrite versus a subtle tweak.
- Switch to a higher-frequency KPI. Test CTR instead of conversion rate — it requires far fewer visitors.
- Increase your daily budget to reach the required sample faster.
- Acknowledge the test is not feasible and make a qualitative decision instead. There is no shame in this — a bad test is worse than no test.
The Confidence Ladder for A/B Tests
Even within a well-planned test, not all intermediate results are equally trustworthy. Here is a framework for how much to trust what you see:
- Under 100 visitors per variant — pure noise. Any difference you see is meaningless. Do not peek.
- 100 to 500 per variant — you might see large directional trends, but they are unstable. If one variant is 3x worse, that is a signal. If it is 20% worse, that is noise.
- 500 to 2,000 per variant — medium confidence. For high-converting pages (10%+), you may be approaching significance. For low-converting ones, keep waiting.
- 2,000 to 5,000 per variant — strong confidence for most common conversion rates.
- 5,000+ per variant — you can detect even small differences with high reliability.
The Most Common Mistakes
After running hundreds of tests across client accounts, these are the patterns we see most often:
1. Peeking and stopping early
You check the results after day 2, see that variant B has a 15% higher conversion rate, and declare victory. The problem is that early results fluctuate wildly. Statistical significance is not something that gradually builds — it can appear and disappear multiple times before settling. This is the "peeking problem", and it inflates your false positive rate dramatically.
Fix: Decide on your sample size before the test starts. Do not look at results until you reach it.
2. Testing too many things at once
You change the headline, the button colour, the hero image, and the CTA text. Variant B converts better — but which change made the difference? You have no idea. And statistically, multiple changes mean more potential for one of them to show a false positive.
Fix: Test one variable at a time. If you must test combinations, use a multivariate test with the appropriate (much larger) sample sizes.
3. Ignoring the base rate
Your page gets 50 conversions a month. You want to detect a 10% improvement. This test will take over a year to reach significance. Do not start it.
Fix: Run the sample size calculation first. If the test is not feasible with your traffic, skip it and make a judgement call instead.
4. Choosing the wrong KPI
You measure ad CTR when what actually matters is cost per conversion. CTR can go up while CPA goes up too — you are attracting more clicks, but worse ones. A "winning" variant might actually be losing you money.
Fix: Choose the KPI that is closest to business value. Conversion rate or revenue per visitor beats CTR almost every time.
5. Never running the numbers
This is the most common one. People launch tests without ever calculating whether they have enough traffic. They check results after an arbitrary period, see a number that looks good, and ship it. This is not testing — it is confirmation bias with a dashboard.
Fix: Before every test, answer: What is my KPI? What is my baseline? What improvement do I want to detect? How many visitors do I need? Can my budget deliver them?
A Practical A/B Testing Checklist
Before you launch any A/B test, go through this checklist:
- Define your KPI. Click rate? Conversion rate? Revenue per visitor? Cost per acquisition? Pick one primary metric.
- Know your baseline. What is the current performance of the metric you are testing?
- Set your minimum detectable effect. What is the smallest change that would matter to your business?
- Calculate required sample size. Use a statistical power calculator. Do not guess.
- Estimate test cost and duration. Sample size times cost per visitor equals test budget. Sample size divided by daily traffic equals duration.
- Decide if it is worth running. If the cost or duration is too high, either test something bolder or skip the test entirely.
- Run the test to completion. Do not peek. Do not stop early unless results are wildly off (3x worse or more).
- Analyse at the pre-set sample size. If the result is significant, ship it. If not, the test is inconclusive — not a failure.
When to Skip A/B Testing Entirely
A/B testing is not always the right tool. Sometimes the answer is obvious and testing would just waste time and money:
- The change is clearly better. Fixing a broken checkout flow, adding a missing CTA, or correcting wrong pricing does not need a test.
- You do not have enough traffic. If statistical significance requires 50,000 visitors and you get 2,000 a month, the test will take over 2 years. Make a decision based on qualitative evidence instead.
- The stakes are too low. Testing whether your footer link colour should be grey or dark grey will not move the needle. Spend your testing budget on high-impact hypotheses.
- You are making a strategic pivot. If you are completely redesigning a page or launching a new product, A/B testing the old versus new is not helpful. You have already decided on the direction — test the details after launch.
The Bottom Line
A/B testing is one of the most powerful tools in digital marketing — but only when done with statistical rigour. The vast majority of tests we see in practice are statistically invalid because they lack the sample size to prove anything.
The fix is simple but requires discipline: start with your KPI, calculate the required sample, estimate the cost, and only run the test if you can afford to do it properly.
A test that cannot reach significance is not a test. It is an expensive way to flip a coin.
Related articles
Marketing Is Shifting From Human-First to Agents-First — Here Is Where We Are
Gartner predicts 15% of daily work decisions will be made by agentic AI by 2028. Marketing is already on this path. We break down the four stages — and where most companies stand today.
Why Ad Creative Experimentation Is the New Competitive Advantage
The companies that experiment the most with ad creatives win. Discover how to build a creative testing machine, leverage failure as data, and drive sustained ROAS.
How Much Should You Invest in Ads? The Statistics Behind Smart Budgeting
Most advertisers set budgets based on a handful of purchases. Discover why this approach is statistically meaningless and learn how to calculate the exact minimum budget your campaigns actually need.