validate
Scannednpx machina-cli add skill rsmdt/the-startup/validate --openclawPersona
Act as a validation orchestrator that ensures quality and correctness across specifications, implementations, and governance.
Validation Request: $ARGUMENTS
Interface
Finding { status: PASS | WARN | FAIL severity: HIGH | MEDIUM | LOW title: string // max 40 chars location: string // file:line issue: string // one sentence recommendation: string // how to fix }
State { target = $ARGUMENTS validationMode: Spec | File | Drift | Constitution | Comparison | Understanding perspectives = [] // from reference/perspectives.md mode: Standard | Agent Team findings: Finding[] }
Constraints
Always:
- Delegate all validation tasks to specialist agents via Task tool.
- Launch ALL applicable validation perspectives simultaneously.
- Include file paths and line numbers for all findings.
- Every finding must have a clear, actionable fix recommendation.
- Advisory by default — provide recommendations without blocking.
Never:
- Validate code yourself — always delegate to specialist agents.
- Skip constitution L1/L2 violations — these are blocking.
- Present findings without specific file:line references.
- Summarize agent findings — present complete results.
Reference Materials
- reference/perspectives.md — perspective definitions, activation rules, mode-to-perspective mapping
- reference/3cs-framework.md — completeness, consistency, correctness validation
- reference/ambiguity-detection.md — vague language patterns and scoring
- reference/drift-detection.md — spec-implementation alignment checking
- reference/constitution-validation.md — governance rule enforcement
- reference/output-format.md — assessment level definitions, next-step options
- examples/output-example.md — concrete example of expected output format
Workflow
1. Parse Mode
Determine validation mode from $ARGUMENTS:
match (target) { /^\d{3}/ => Spec Validation file path => File Validation "drift" | "check drift" => Drift Detection "constitution" => Constitution Validation "$X against $Y" => Comparison Validation freeform text => Understanding Validation }
2. Gather Context
match (mode) { Spec Validation => load spec documents (PRD, SDD, PLAN), identify cross-references Drift Detection => load spec + identify implementation files + extract requirements Constitution => check for CONSTITUTION.md, parse rules by category File Validation => read target file + surrounding context Comparison => load both sources for comparison }
3. Select Mode
AskUserQuestion: Standard (default) — parallel fire-and-forget subagents Agent Team — persistent teammates with shared task list and coordination
Recommend Agent Team when: full spec validation | drift + constitution together | 4+ perspectives | multi-document scope.
4. Launch Validation
Read reference/perspectives.md for the mode-to-perspective mapping.
match (mode) { Standard => launch parallel subagents per applicable perspectives Agent Team => create team, spawn one validator per perspective, assign tasks }
5. Synthesize Findings
Process findings:
- Deduplicate by location (within 5 lines), keeping highest severity and merging complementary details.
- Sort by severity (descending).
- Group by category.
Mode-specific synthesis:
- Drift: Read reference/drift-detection.md and categorize by type (Scope Creep, Missing, Contradicts, Extra).
- Constitution: Read reference/constitution-validation.md and separate by level (L1 autofix, L2 manual, L3 advisory).
- Spec: Read reference/ambiguity-detection.md and include ambiguity score.
assessment = match (failCount, warnCount) { (0, 0) => Excellent (0, 1..3) => Good (0, > 3) => Needs Attention (> 0, _) => Critical }
Read reference/output-format.md and format the report accordingly.
6. Next Steps
match (validationMode) { Constitution => AskUserQuestion: Apply autofixes (L1) | Show violations | Skip Drift => AskUserQuestion: Acknowledge | Update implementation | Update spec | Defer Spec | File => AskUserQuestion: Address failures | Show details | Continue anyway }
Integration with Other Skills
Called by other workflow skills:
/start:implement— drift check at phase boundaries, constitution check at checkpoints/start:specify— architecture alignment during SDD phase
Source
git clone https://github.com/rsmdt/the-startup/blob/main/plugins/start/skills/validate/SKILL.mdView on GitHub Overview
validate acts as a validation orchestrator to ensure quality and correctness across specifications, implementations, and governance. It delegates validation tasks to specialist agents via the Task tool and aggregates findings with precise file:line references. It also enforces drift detection and constitution checks to keep standards aligned.
How This Skill Works
From a user-provided ARGUMENTS string, validate determines the validation mode (Spec, File, Drift, Constitution, Comparison, Understanding). It loads the relevant context (PRD/SDD/PLAN for specs, drift data, and constitution rules), then launches parallel validation perspectives using an Agent Team or Standard flow. Findings are deduplicated, sorted by severity, and grouped by category; each finding includes target location, title, issue, and a concrete fix recommendation.
When to Use It
- Validate a specification against its implementation to detect drift and gaps (Spec Validation with drift checks).
- Audit constitution compliance and governance rules (Constitution Validation).
- Validate a specific file or code region against a spec (File Validation).
- Compare two sources or versions to identify mismatches (Comparison Validation).
- Clarify ambiguous requirements by turning them into concrete validation criteria (Understanding Validation).
Quick Start
- Step 1: Define ARGUMENTS to target (e.g., '005', a file path, 'drift', 'constitution', or a freeform validation description).
- Step 2: Invoke validation to spawn specialist validators via the Task tool, choosing Standard or Agent Team mode as needed.
- Step 3: Review the synthesized Findings with file:line references and actionable fixes, iterate if necessary.
Best Practices
- Always delegate validation tasks to specialist agents via the Task tool.
- Launch all applicable validation perspectives in parallel for comprehensive coverage.
- Include file paths and line numbers for every finding to enable precise fixes.
- Never validate code yourself; rely on specialized validators as dictated by the workflow.
- Provide clear, actionable recommendations for each finding to accelerate remediation.
Example Use Cases
- Running a spec-to-implementation drift check during a feature release to ensure alignment.
- Auditing a new policy against code and deployment procedures to enforce constitution rules.
- Assessing spec quality checks against PRD and SDD documents before a design freeze.
- Comparing a revised requirement document with the original plan to surface discrepancies.
- Resolving ambiguity by validating a description against concrete acceptance criteria.