content-experimentation-best-practices
npx machina-cli add skill sanity-io/agent-toolkit/content-experimentation-best-practices --openclawContent Experimentation Best Practices
Principles and patterns for running effective content experiments to improve conversion rates, engagement, and user experience.
When to Apply
Reference these guidelines when:
- Setting up A/B or multivariate testing infrastructure
- Designing experiments for content changes
- Analyzing and interpreting test results
- Building CMS integrations for experimentation
- Deciding what to test and how
Core Concepts
A/B Testing
Comparing two variants (A vs B) to determine which performs better.
Multivariate Testing
Testing multiple variables simultaneously to find optimal combinations.
Statistical Significance
The confidence level that results aren't due to random chance.
Experimentation Culture
Making decisions based on data rather than opinions (HiPPO avoidance).
Resources
See resources/ for detailed guidance:
resources/experiment-design.md— Hypothesis framework, metrics, sample size, and what to testresources/statistical-foundations.md— p-values, confidence intervals, power analysis, Bayesian methodsresources/cms-integration.md— CMS-managed variants, field-level variants, external platformsresources/common-pitfalls.md— 17 common mistakes across statistics, design, execution, and interpretation
Source
git clone https://github.com/sanity-io/agent-toolkit/blob/main/skills/content-experimentation-best-practices/SKILL.mdView on GitHub Overview
This skill codifies principles for running data-driven content experiments to improve conversions, engagement, and UX. It covers choosing testing approaches (A/B and multivariate), interpreting results, and building CMS-supported experimentation workflows.
How This Skill Works
Begin with a clear hypothesis and defined metrics. Choose a testing method (A/B or multivariate) based on the content change scope, then run the experiment with appropriate sample size and significance checks. Use the analysis to decide on content changes and leverage CMS integrations to manage variants and automate deployment.
When to Use It
- Setting up A/B or multivariate testing infrastructure
- Designing experiments for content changes
- Analyzing and interpreting test results
- Building CMS integrations for experimentation
- Deciding what to test and how
Quick Start
- Step 1: Define hypothesis, metrics, and success criteria.
- Step 2: Design variants and configure CMS/infrastructure to deliver them.
- Step 3: Run the experiment, monitor results, and analyze significance to act on findings.
Best Practices
- Define clear hypotheses, metrics, and success criteria
- Choose the right test type (A/B vs multivariate) for the change
- Plan for adequate sample size, power, and statistical significance
- Integrate CMS features for variant management and CMS-driven experimentation
- Be mindful of common pitfalls and HiPPO bias; use data-first decision making
Example Use Cases
- Landing page headline A/B test to boost CTR
- Hero section multivariate test (image, headline, CTA) on a product page
- CMS-driven variant routing to test field-level changes during a campaign
- Experimentation pipeline to monitor p-values during a product launch
- Post-test analysis to decide full deployment or rollback