Bayesian A/B Testing: Decision-Making with Posterior Distributions
This paper outlines a practical Bayesian framing for A/B tests, focusing on posterior probability and expected loss to guide go/no-go decisions in product experiments.
Why Bayesian A/B Testing?
Frequentist A/B tests are useful, but decision-making often needs answers like:
- What is the probability that variant B beats variant A?
- How large is the uplift likely to be?
- What is the risk of shipping a change with small or negative effect?
A Bayesian approach provides these probabilities directly by modeling a posterior distribution over conversion rates.
Model Setup
Assume Bernoulli conversions with Beta priors:
After observing conversions out of exposures, the posterior is:
Decision Metrics
Useful metrics derived from the posterior include:
- , the probability of improvement.
- Expected uplift, .
- Expected loss if selecting the wrong variant.
These map well to product decisions where the cost of a wrong launch is explicit.
Practical Guidance
- Start with weakly informative priors when historical data is limited.
- Favor decisions based on expected loss, not only on tail probabilities.
- Report uncertainty intervals to communicate the range of plausible outcomes.
Conclusion
Bayesian A/B testing complements product decision-making by turning uncertain outcomes into probabilistic guidance. The framework scales from quick experiments to complex multi-variant testing.