Get the FREE Ultimate OpenClaw Setup Guide →

data-validate

Scanned
npx machina-cli add skill kevinlin/cowork-z/data-validate --openclaw
Files (1)
SKILL.md
5.8 KB

/validate - Validate Analysis Before Sharing

If you see unfamiliar placeholders or need to check which tools are connected, see CONNECTORS.md.

Review an analysis for accuracy, methodology, and potential biases before sharing with stakeholders. Generates a confidence assessment and improvement suggestions.

Usage

/validate <analysis to review>

The analysis can be:

  • A document or report in the conversation
  • A file (markdown, notebook, spreadsheet)
  • SQL queries and their results
  • Charts and their underlying data
  • A description of methodology and findings

Workflow

1. Review Methodology and Assumptions

Examine:

  • Question framing: Is the analysis answering the right question? Could the question be interpreted differently?
  • Data selection: Are the right tables/datasets being used? Is the time range appropriate?
  • Population definition: Is the analysis population correctly defined? Are there unintended exclusions?
  • Metric definitions: Are metrics defined clearly and consistently? Do they match how stakeholders understand them?
  • Baseline and comparison: Is the comparison fair? Are time periods, cohort sizes, and contexts comparable?

2. Check for Common Analytical Errors

Systematically review for:

Data completeness:

  • Missing data that could skew results (e.g., nulls in key fields, missing time periods)
  • Data freshness issues (is the most recent data actually complete or still loading?)
  • Survivorship bias (are you only looking at entities that "survived" to the analysis date?)

Statistical issues:

  • Simpson's paradox (trend reverses when data is aggregated vs. segmented)
  • Correlation presented as causation without supporting evidence
  • Small sample sizes leading to unreliable conclusions
  • Outliers disproportionately affecting averages (should medians be used instead?)
  • Multiple testing / cherry-picking significant results

Aggregation errors:

  • Double-counting from improper joins (many-to-many explosions)
  • Incorrect denominators in rate calculations
  • Mixing granularity levels (e.g., user-level metrics averaged with account-level)
  • Revenue recognized vs. billed vs. collected confusion

Time-related issues:

  • Seasonality not accounted for in comparisons
  • Incomplete periods included in averages (e.g., partial month compared to full months)
  • Timezone inconsistencies between data sources
  • Look-ahead bias (using future information to explain past events)

Selection and scope:

  • Cherry-picked time ranges that favor a particular narrative
  • Excluded segments without justification
  • Changing definitions mid-analysis

3. Verify Calculations and Aggregations

Where possible, spot-check:

  • Recalculate a few key numbers independently
  • Verify that subtotals sum to totals
  • Check that percentages sum to 100% (or close to it) where expected
  • Confirm that YoY/MoM comparisons use the correct base periods
  • Validate that filters are applied consistently across all metrics

4. Assess Visualizations

If the analysis includes charts:

  • Do axes start at appropriate values (zero for bar charts)?
  • Are scales consistent across comparison charts?
  • Do chart titles accurately describe what's shown?
  • Could the visualization mislead a quick reader?
  • Are there truncated axes, inconsistent intervals, or 3D effects that distort perception?

5. Evaluate Narrative and Conclusions

Review whether:

  • Conclusions are supported by the data shown
  • Alternative explanations are acknowledged
  • Uncertainty is communicated appropriately
  • Recommendations follow logically from findings
  • The level of confidence matches the strength of evidence

6. Suggest Improvements

Provide specific, actionable suggestions:

  • Additional analyses that would strengthen the conclusions
  • Caveats or limitations that should be noted
  • Better visualizations or framings for key points
  • Missing context that stakeholders would want

7. Generate Confidence Assessment

Rate the analysis on a 3-level scale:

Ready to share -- Analysis is methodologically sound, calculations verified, caveats noted. Minor suggestions for improvement but nothing blocking.

Share with noted caveats -- Analysis is largely correct but has specific limitations or assumptions that must be communicated to stakeholders. List the required caveats.

Needs revision -- Found specific errors, methodological issues, or missing analyses that should be addressed before sharing. List the required changes with priority order.

Output Format

## Validation Report

### Overall Assessment: [Ready to share | Share with caveats | Needs revision]

### Methodology Review
[Findings about approach, data selection, definitions]

### Issues Found
1. [Severity: High/Medium/Low] [Issue description and impact]
2. ...

### Calculation Spot-Checks
- [Metric]: [Verified / Discrepancy found]
- ...

### Visualization Review
[Any issues with charts or visual presentation]

### Suggested Improvements
1. [Improvement and why it matters]
2. ...

### Required Caveats for Stakeholders
- [Caveat that must be communicated]
- ...

Examples

/validate Review this quarterly revenue analysis before I send it to the exec team: [analysis]
/validate Check my churn analysis -- I'm comparing Q4 churn rates to Q3 but Q4 has a shorter measurement window
/validate Here's a SQL query and its results for our conversion funnel. Does the logic look right? [query + results]

Tips

  • Run /validate before any high-stakes presentation or decision
  • Even quick analyses benefit from a sanity check -- it takes a minute and can save your credibility
  • If the validation finds issues, fix them and re-validate
  • Share the validation output alongside your analysis to build stakeholder confidence

Source

git clone https://github.com/kevinlin/cowork-z/blob/main/src-tauri/resources/skill-templates/data-validate/SKILL.mdView on GitHub

Overview

data-validate helps QA an analysis before sharing with stakeholders. It evaluates methodology, data selection, population definitions, metrics, and potential biases, and it generates a confidence assessment with actionable improvement suggestions.

How This Skill Works

Provide the analysis to review via /validate. The tool systematically examines methodology, assumptions, data selection, and metric definitions, flags common analytical errors, verifies calculations and aggregations, and checks visualizations and narrative coherence. It returns a confidence assessment and concrete improvement suggestions.

When to Use It

  • Review a document or report before sharing with stakeholders
  • Validate a SQL query, notebook, or dashboard before release
  • Audit charts and figures to prevent misleading visuals
  • Reassess methodology and population definitions when data sources change
  • Assess bias, uncertainty, and alternative explanations in findings

Quick Start

  1. Step 1: Prepare the analysis (document, file, notebook, SQL, or chart) and run /validate <analysis to review>
  2. Step 2: Review the confidence assessment and suggested improvements returned by data-validate
  3. Step 3: Implement the recommended changes and re-run /validate to confirm improvements

Best Practices

  • Review methodology and assumptions: check question framing, data selection, population definition, metric definitions, and baseline/comparison
  • Check for common analytical errors: data completeness, statistical pitfalls, aggregation issues, time-related biases, and scope exclusions
  • Verify calculations and aggregations: recalculate key numbers, ensure totals/subtotals align, and verify base periods for YoY/MoM
  • Assess visualizations: ensure axes, scales, and titles are accurate and not misleading; watch for truncated axes
  • Evaluate narrative and conclusions: confirm conclusions align with data, acknowledge uncertainty, and present clearly stated limitations

Example Use Cases

  • QA a campaign ROI analysis to ensure data freshness and correct denominator definitions before presenting to marketing leadership
  • Audit a sales funnel study for survivorship bias and incorrect time windows prior to a quarterly review
  • Validate a finance forecast with look-ahead bias checks and ensure proper baseline periods are used
  • Review a product metrics dashboard for data gaps and inconsistent aggregations across cohorts
  • Assess a research report’s methodology and limitations before publication to stakeholders

Frequently Asked Questions

Add this skill to your agents
Sponsor this space

Reach thousands of developers