Measure Experiment Design
Scannednpx machina-cli add skill product-on-purpose/pm-skills/measure-experiment-design --openclawname: measure-experiment-design description: Designs an A/B test or experiment with clear hypothesis, variants, success metrics, sample size, and duration. Use when planning experiments to validate product changes or test hypotheses. phase: measure version: "2.0.0" updated: 2026-01-26 license: Apache-2.0 metadata: category: validation frameworks: [triple-diamond, lean-startup, design-thinking] author: product-on-purpose
Experiment Design
An experiment design document defines all parameters needed to run a rigorous A/B test or controlled experiment. It ensures the team aligns on what you're testing, how you'll measure success, and how long to run the test before drawing conclusions. Good experiment design prevents common pitfalls: underpowered tests, unclear success criteria, and decisions based on noise rather than signal.
When to Use
- Before launching an A/B test to validate a product change
- When testing a hypothesis that requires quantitative validation
- After solution design to validate assumptions before full rollout
- When stakeholders want data-driven evidence for a decision
- To establish a culture of experimentation and learning
Instructions
When asked to design an experiment, follow these steps:
-
Articulate the Hypothesis Write a clear, testable hypothesis in the format: "We believe [change] for [users] will [outcome] as measured by [metric]." One hypothesis per experiment — if you're testing multiple things, run multiple experiments.
-
Define the Variants Describe the control (current experience) and treatment (new experience) in sufficient detail. Include screenshots, mockups, or precise descriptions so anyone can understand what users will see.
-
Choose Primary and Secondary Metrics Select one primary metric that will determine success or failure. Add 2-3 secondary metrics to understand the broader impact. Include guardrail metrics to catch unintended negative effects.
-
Calculate Sample Size Determine how many users you need per variant to detect your minimum detectable effect (MDE) with statistical significance. Specify your significance level (typically 0.05) and power (typically 0.80).
-
Estimate Duration Based on sample size and available traffic, calculate how long the experiment needs to run. Account for weekly patterns — avoid ending mid-week if behavior varies by day.
-
Define Targeting and Allocation Specify which users are eligible for the experiment and how traffic is split between variants. Document any exclusions (e.g., employees, specific segments).
-
Set Success Criteria Define upfront what constitutes a win, a loss, or an inconclusive result. This prevents post-hoc rationalization and moving goalposts.
-
Document Risks and Mitigations Identify what could go wrong and how you'll detect/address it. Include monitoring plans and rollback criteria.
Output Format
Use the template in references/TEMPLATE.md to structure the output.
Quality Checklist
Before finalizing, verify:
- Hypothesis is falsifiable and specific
- Only one primary metric is defined
- Sample size calculation is documented with assumptions
- Duration accounts for traffic patterns and statistical requirements
- Success criteria are defined before the experiment starts
- Guardrail metrics protect against unintended harm
Examples
See references/EXAMPLE.md for a completed example.
Source
git clone https://github.com/product-on-purpose/pm-skills/blob/main/skills/measure-experiment-design/SKILL.mdView on GitHub Overview
Measure Experiment Design guides you through planning a rigorous A/B test or controlled experiment. It helps teams align on what to test, how to measure success, and how long to run, reducing underpowered tests and noisy results.
How This Skill Works
Technically, you document a test plan with a falsifiable hypothesis, clearly described control and treatment variants, and a primary metric with guardrails. You also calculate the required sample size for the minimum detectable effect with a chosen alpha and power, estimate the test duration, and define targeting and allocation. Finally, you set predefined success criteria, document risks and mitigations, and prepare monitoring and rollback strategies.
When to Use It
- Before launching an A/B test to validate a product change
- When testing a hypothesis that requires quantitative validation
- After solution design to validate assumptions before full rollout
- When stakeholders want data-driven evidence for a decision
- To establish a culture of experimentation and learning
Quick Start
- Step 1: Articulate a falsifiable hypothesis in the format We believe [change] for [users] will [outcome] as measured by [metric].
- Step 2: Define the control and treatment variants with clear descriptions or visuals.
- Step 3: Compute sample size for the minimum detectable effect, set alpha and power, estimate duration, and lock in targeting, success criteria, and risk plan.
Best Practices
- Hypothesis is falsifiable and specific
- Only one primary metric defined
- Sample size calculated with clear assumptions and MDE
- Duration accounts for traffic patterns and statistical requirements
- Guardrail metrics, monitoring, and rollback plans defined upfront
Example Use Cases
- Validate a pricing change and its impact on conversion
- Test a redesigned onboarding flow in a target segment
- Evaluate a new search ranking variant for engagement
- Roll out a feature flag to a subset of users and compare
- Assess checkout optimization effects on completion rate and revenue