test-review
npx machina-cli add skill athola/claude-night-market/test-review --openclawTable of Contents
- Quick Start
- When to Use
- Required TodoWrite Items
- Progressive Loading
- Workflow
- Step 1: Detect Languages (
test-review:languages-detected) - Step 2: Inventory Coverage (
test-review:coverage-inventoried) - Step 3: Assess Scenario Quality (
test-review:scenario-quality) - Step 4: Plan Remediation (
test-review:gap-remediation) - Step 5: Log Evidence (
test-review:evidence-logged) - Test Quality Checklist (Condensed)
- Output Format
- Summary
- Framework Detection
- Coverage Analysis
- Quality Issues
- Remediation Plan
- Recommendation
- Integration Notes
- Exit Criteria
Test Review Workflow
Evaluate and improve test suites with TDD/BDD rigor.
Quick Start
/test-review
Verification: Run pytest -v to verify tests pass.
When To Use
- Reviewing test suite quality
- Analyzing coverage gaps
- Before major releases
- After test failures
- Planning test improvements
When NOT To Use
- Writing new tests - use parseltongue:python-testing
- Updating existing tests - use sanctum:test-updates
Required TodoWrite Items
test-review:languages-detectedtest-review:coverage-inventoriedtest-review:scenario-qualitytest-review:gap-remediationtest-review:evidence-logged
Progressive Loading
Load modules as needed based on review depth:
- Basic review: Core workflow (this file)
- Framework detection: Load
modules/framework-detection.md - Coverage analysis: Load
modules/coverage-analysis.md - Quality assessment: Load
modules/scenario-quality.md - Remediation planning: Load
modules/remediation-planning.md
Workflow
Step 1: Detect Languages (test-review:languages-detected)
Identify testing frameworks and version constraints.
→ See: modules/framework-detection.md
Quick check:
find . -maxdepth 2 -name "Cargo.toml" -o -name "pyproject.toml" -o -name "package.json" -o -name "go.mod"
Verification: Run the command with --help flag to verify availability.
Step 2: Inventory Coverage (test-review:coverage-inventoried)
Run coverage tools and identify gaps.
→ See: modules/coverage-analysis.md
Quick check:
git diff --name-only | rg 'tests|spec|feature'
Verification: Run pytest -v to verify tests pass.
Step 3: Assess Scenario Quality (test-review:scenario-quality)
Evaluate test quality using BDD patterns and assertion checks.
→ See: modules/scenario-quality.md
Focus on:
- Given/When/Then clarity
- Assertion specificity
- Anti-patterns (dead waits, mocking internals, repeated boilerplate)
Step 4: Plan Remediation (test-review:gap-remediation)
Create concrete improvement plan with owners and dates.
→ See: modules/remediation-planning.md
Step 5: Log Evidence (test-review:evidence-logged)
Record executed commands, outputs, and recommendations.
→ See: imbue:evidence-logging
Test Quality Checklist (Condensed)
- Clear test structure (Arrange-Act-Assert)
- Critical paths covered (auth, validation, errors)
- Specific assertions with context
- No flaky tests (dead waits, order dependencies)
- Reusable fixtures/factories
Output Format
## Summary
[Brief assessment]
## Framework Detection
- Languages: [list] | Frameworks: [list] | Versions: [constraints]
## Coverage Analysis
- Overall: X% | Critical: X% | Gaps: [list]
## Quality Issues
[Q1] [Issue] - Location - Fix
## Remediation Plan
1. [Action] - Owner - Date
## Recommendation
Approve / Approve with actions / Block
Verification: Run the command with --help flag to verify availability.
Integration Notes
- Use
imbue:evidence-loggingfor reproducible evidence capture - Reference
imbue:diff-analysisfor risk assessment - Format output using
imbue:structured-outputpatterns
Exit Criteria
- Frameworks detected and documented
- Coverage analyzed and gaps identified
- Scenario quality assessed
- Remediation plan created with owners and dates
- Evidence logged with citations
Troubleshooting
Common Issues
Tests not discovered
Ensure test files match pattern test_*.py or *_test.py. Run pytest --collect-only to verify.
Import errors
Check that the module being tested is in PYTHONPATH or install with pip install -e .
Async tests failing
Install pytest-asyncio and decorate test functions with @pytest.mark.asyncio
Source
git clone https://github.com/athola/claude-night-market/blob/master/plugins/pensive/skills/test-review/SKILL.mdView on GitHub Overview
Test Review evaluates test suites for coverage gaps and quality, guiding audits, remediation planning, and evidence logging. It aligns with TDD/BDD practices and uses a progressive loading workflow to scope analysis. Do not use for writing new tests or updating existing tests, per the documented exclusions.
How This Skill Works
The workflow starts by identifying test languages/frameworks (test-review:languages-detected), then inventories coverage (test-review:coverage-inventoried), and finally assesses scenario quality (test-review:scenario-quality) to plan remediation (test-review:gap-remediation) and log evidence (test-review:evidence-logged). Modules are loaded progressively (framework-detection, coverage-analysis, scenario-quality, remediation-planning, content-assertion-quality) to control depth.
When to Use It
- Reviewing test suite quality
- Analyzing coverage gaps
- Before major releases
- After test failures
- Planning test improvements
Quick Start
- Step 1: Run /test-review to initialize the workflow.
- Step 2: Verify test execution with pytest -v.
- Step 3: Progress through Step 1-5: languages-detected, coverage-inventoried, scenario-quality, gap-remediation, evidence-logged.
Best Practices
- Start with languages/frameworks detection to tailor analysis.
- Inventory coverage before proposing remediation.
- Evaluate Given/When/Then clarity and assertion specificity.
- Log evidence and preserve artifacts for traceability.
- Use progressive loading to manage review depth and scope.
Example Use Cases
- Audit a Python pytest suite to identify API coverage gaps and flaky tests.
- Assess feature scenarios for Given/When/Then clarity and replace vague steps.
- Plan remediation for missing test cases and align with gap-finder outputs.
- Leverage coverage-analyzer and scenario-evaluator to quantify quality deficits.
- Conduct pre-release reviews to ensure coverage improvements prior to release.
Frequently Asked Questions
Related Skills
precommit-setup
athola/claude-night-market
Configure three-layer pre-commit system with linting, type checking, and testing hooks. Use for quality gate setup and code standards. Skip if pre-commit is optimally configured.
ab-test-setup
coreyhaines31/marketingskills
When the user wants to plan, design, or implement an A/B test or experiment. Also use when the user mentions "A/B test," "split test," "experiment," "test this change," "variant copy," "multivariate test," "hypothesis," "should I test this," "which version is better," "test two versions," "statistical significance," or "how long should I run this test." Use this whenever someone is comparing two approaches and wants to measure which performs better. For tracking implementation, see analytics-tracking. For page-level conversion optimization, see page-cro.
Playwright Browser Automation
jpulido240-svg/playwright-skill
Complete browser automation with Playwright. Auto-detects dev servers, writes clean test scripts to /tmp. Test pages, fill forms, take screenshots, check responsive design, validate UX, test login flows, check links, automate any browser task. Use when user wants to test websites, automate browser interactions, validate web functionality, or perform any browser-based testing.
python-testing
athola/claude-night-market
'Consult this skill for Python testing implementation and patterns. Use
workflow-setup
athola/claude-night-market
Configure GitHub Actions CI/CD workflows for automated testing, linting, and deployment. Use for CI/CD setup and quality automation. Skip if CI/CD configured or using different platform.
RubyCritic Code Quality Analysis
esparkman/claude-rubycritic-skill
Analyze Ruby and Rails code quality with RubyCritic. Identifies code smells, complexity issues, and refactoring opportunities. Provides detailed metrics, scores files A-F, compares branches, and prioritizes high-churn problem areas. Use when analyzing Ruby code quality, reviewing PRs, or identifying technical debt.