shipkit-dev-review
Scannednpx machina-cli add skill stefan-stepzero/shipkit/shipkit-dev-review --openclawshipkit-dev-review - Framework Design Review
Purpose: Catch design issues in framework changes that structural validation misses
What it does:
- Reviews recent changes or specific skills for design quality
- Validates Skill Value Test compliance
- Checks pattern consistency across the framework
- Verifies Claude Code compatibility
- Produces a review report with actionable findings
How it differs from existing quality tools:
| Tool | What it checks |
|---|---|
shipkit-framework-integrity | Structural + consistency: frontmatter, 7-file integration, manifest sync, broken refs, installer |
shipkit-dev-review | Design: is this the RIGHT thing built the RIGHT way? |
When to Invoke
User says:
- "Review this change"
- "Is this skill well designed?"
- "Code review the framework changes"
- "Check quality before release"
- "Review what we just built"
Automated trigger:
- After implementing framework changes
- Before running
/shipkit-dev-release - As the reviewer role in a
/shipkit-dev-team
Prerequisites
Required:
- Changes to review (recent commits, a branch, or specific files)
Helpful context:
.claude/specs/{feature}.json— spec the changes implementCLAUDE.md— Skill Value Test, framework rulesdocs/development/SKILL-QUALITY-AND-PATTERNS.md— quality standards
Process
Step 1: Identify Scope
Based on argument:
--scope recent— Review uncommitted changes + last commit (default)--scope branch— Review all commits on current branch vs main--scope skill:<name>— Review a specific skill in depth--scope all— Review entire framework (expensive, use sparingly)
# For recent
git diff HEAD~1 --name-only
git diff --name-only
# For branch
git log main..HEAD --name-only --pretty=format:""
Step 2: Categorize Changes
Group changed files by component type:
- Skills —
install/skills/shipkit-*/SKILL.md - Local skills —
.claude/skills/shipkit-*/SKILL.md - Agents —
install/agents/shipkit-*.md - Hooks —
install/shared/hooks/*.py - Config —
install/settings/,install/profiles/ - Docs —
docs/generated/,README.md - Rules —
install/rules/ - Templates —
install/claude-md/
Step 3: Design Review — Skills
For each changed/new skill, check:
3.1 Value Test
- Does this skill force explicit human decisions? Which ones?
- Does this skill create persistence Claude lacks? What persists?
- If neither → FLAG: "Skill may be redundant"
3.2 Scope Discipline
- Does the skill do ONE thing well?
- Is the process section < 10 steps? (More = over-engineered)
- Does it avoid implementing what Claude does naturally?
- Are there steps that could be removed without losing value?
3.3 Context Awareness
- Does it read relevant
.shipkit/context before acting? - Does it check for existing decisions instead of assuming defaults?
- Are context file paths correct and existing?
3.4 Integration Quality
- "Before/After" integrations make sense causally (not just lists)
- Cross-references are bidirectional
- No circular dependencies
3.5 Output Quality
- Outputs are structured (JSON for data, MD for narrative)
- JSON outputs follow artifact convention
- Outputs are consumed by a downstream skill (not dead-end data)
3.6 Pattern Consistency
- Follows same frontmatter pattern as peer skills
- Section structure matches framework conventions
- Naming matches conventions (CLAUDE.md naming table)
Step 4: Design Review — Agents
For each changed/new agent:
- Role is distinct from existing agents (no overlap)
- Model selection matches role complexity (Haiku for simple, Sonnet for standard, Opus for reasoning-heavy)
-
permissionModeis appropriate - Agent prompt doesn't duplicate skill instructions
- If
memoryfield used, scoping is appropriate
Step 5: Design Review — Hooks
For each changed/new hook:
- Hook event is correct for the trigger
- Exit codes follow convention (0 = allow, 2 = block with feedback)
- Error handling uses silent failure pattern (never crashes)
- Timeout is appropriate for what the hook does
- State files use
.local.naming (gitignored) - Hook is idempotent (safe to fire multiple times)
Step 6: Design Review — Cross-Cutting
Check framework-wide concerns:
- No duplicate functionality between skills
- Skill count in README/overview matches actual count
- Settings permissions cover all new skills
- Manifest entries match disk reality
- No secrets or credentials in committed files
Step 7: CC Compatibility Check
Against current Claude Code version:
- All frontmatter fields are recognized
- No deprecated patterns used
- Hook events are valid
- Settings schema is current
Step 8: Write Review Report
Present findings inline (not a JSON file — reviews are transient):
## Framework Design Review
**Scope**: {what was reviewed}
**Date**: {date}
### Findings
#### BLOCK: {title}
{description}
**File**: {path}:{line}
**Fix**: {concrete action}
#### WARN: {title}
{description}
**File**: {path}:{line}
**Suggestion**: {what to consider}
#### NOTE: {title}
{description}
### Summary
- Blockers: {N}
- Warnings: {N}
- Notes: {N}
- **Verdict**: {PASS / PASS WITH WARNINGS / BLOCK}
If --strict mode: Warnings are promoted to blockers.
Severity Levels
| Level | Meaning | Action |
|---|---|---|
| BLOCK | Must fix before merge/release | Stop and fix |
| WARN | Should fix, but not blocking | Fix or document why not |
| NOTE | Observation, no action needed | Awareness only |
When This Skill Integrates with Others
Before This Skill
- Implementation — changes must exist to review
/shipkit-framework-integrity— structural checks should pass first- Why: Don't waste design review time on structurally broken skills
After This Skill
/shipkit-dev-release— Release after review passes- Trigger: Review verdict is PASS
- Why: Only release reviewed code
- Fix and re-review — if blockers found
Context Files This Skill Reads
.claude/specs/{feature}.json— Spec being implemented (if available)CLAUDE.md— Skill Value Test, framework rulesdocs/development/SKILL-QUALITY-AND-PATTERNS.md— Quality standards- All changed files in scope
Context Files This Skill Writes
- None — review output is presented inline, not persisted
Source
git clone https://github.com/stefan-stepzero/shipkit/blob/main/.claude/skills/shipkit-dev-review/SKILL.mdView on GitHub Overview
shipkit-dev-review evaluates design quality in framework changes, not just structure. It checks Skill Value Test compliance, pattern consistency, and Claude Code compatibility, delivering actionable findings before release.
How This Skill Works
The tool analyzes the defined scope (--scope), categorizes changed files by component, and for each changed skill runs design checks across VALUE TEST, SCOPE DISCIPLINE, CONTEXT AWARENESS, INTEGRATION QUALITY, OUTPUT QUALITY, and PATTERN CONSISTENCY. Outputs are produced in a structured JSON/MD format and aligned with artifact conventions for downstream consumption.
When to Use It
- Review this change
- Is this skill well designed?
- Code review the framework changes
- Check quality before release
- Review what we just built
Quick Start
- Step 1: Run with --scope recent|branch|skill:<name>|all to define the review area
- Step 2: Categorize changed files by Skills, Local skills, Agents, Hooks, Config, Docs, Rules, Templates
- Step 3: Run the design review to produce a structured findings report and recommendations
Best Practices
- For each changed skill, verify explicit human decisions are required (VALUE TEST).
- Keep the scope to ONE clear outcome per skill to avoid over-engineering.
- Ensure context awareness by reading relevant .shipkit context and existing decisions.
- Confirm cross-references and bidirectional links; avoid circular dependencies.
- Ensure outputs follow artifact conventions and are usable by downstream skills.
Example Use Cases
- Assess a recent skill addition for design quality and alignment with CLAUDE.md.
- Validate a feature branch's framework changes before merging to main.
- Verify Skill Value Test alignment against the Skill's spec and docs.
- Check Claude Code compatibility and consistent pattern usage across skills.
- Generate a findings report and action list for release readiness.