Experimentation Patterns
What are Experimentation Patterns?
Experimentation patterns are systematic approaches to testing hypotheses about user behaviour, feature effectiveness, and product improvements through controlled experiments (e.g. A/B tests). They help you validate assumptions and make decisions based on evidence rather than opinion.
Use it when: you have a clear hypothesis about a change (copy, layout, flow) and want to learn from real user behaviour before committing fully.
Copy/paste template
- Hypothesis: [If we X, then Y will improve because Z.]
- Primary metric: [what you’re optimising, e.g. sign-up rate]
- Secondary metrics: [guardrails, e.g. retention, time on task]
- Variants: [control vs treatment(s), clearly defined]
- Sample size / duration: [enough for significance; avoid stopping early]
- Success criteria: [what would make you ship or roll back]
Why Experimentation Patterns matter
- Replace guesswork with evidence so you ship what actually works.
- Reduce risk by testing with a subset of users before full rollout.
- Build a culture of learning and data-informed decision making.
- Focus effort on changes that move the needle on key metrics.
- Create a shared language for "we believe X; let’s test it."
What a good experiment includes
Checklist
- [ ] Clear, testable hypothesis (if X, then Y; measurable).
- [ ] Single variable (or controlled multivariate design) so you know what drove the result.
- [ ] Adequate sample size and duration for statistical reliability.
- [ ] Primary and guardrail metrics so you don’t optimise one thing at the expense of others.
- [ ] Pre-registered plan (what you’re testing, when you’ll look, what you’ll do) to avoid cherry-picking.
Common formats
- A/B test: one control, one treatment; compare on a primary metric.
- Multivariate / MVT: test multiple factors; use when you have traffic and need to understand interactions.
Examples
Example (the realistic one)
Hypothesis: "If we shorten the sign-up form from 8 fields to 4, completion rate will increase without hurting quality." They run an A/B test, primary metric = completion rate, guardrail = lead quality (e.g. conversion to paid). After a pre-set sample size, the short form wins on completion and quality is unchanged. They ship the short form.
Common pitfalls
- Testing without a hypothesis: running tests to "see what happens". → Do this instead: write a hypothesis and success criteria before you start.
- Stopping early or peeking: deciding based on incomplete data. → Do this instead: pre-define sample size/duration and stick to it, or use sequential testing methods correctly.
- Changing too many things: you can’t tell what caused the result. → Do this instead: test one change (or a clearly defined set) per experiment.
- Ignoring guardrail metrics: optimising one metric while harming another. → Do this instead: define primary and guardrail metrics; don’t ship if guardrails regress.
Experimentation Patterns vs. related concepts
- Experimentation vs A/B testing: A/B testing is one experimentation pattern; experimentation patterns also include design of experiments, metrics, and decision rules.
- Experimentation vs user research: Experiments measure behaviour at scale; user research (e.g. interviews) explains why. Use both: research to form hypotheses, experiments to test them.
- Experimentation vs gut feel: Experimentation uses data and structure; gut feel is unvalidated. Use experiments when the stakes or uncertainty are high.
Related terms
Next step
If you’re new to running experiments, start with A/B testing and define one clear hypothesis and metric. If you’re forming hypotheses from user insight, use user research and usability testing.