Get the FREE Ultimate OpenClaw Setup Guide →

primr-qa

Scanned
npx machina-cli add skill blisspixel/primr/primr-qa --openclaw
Files (1)
SKILL.md
4.5 KB

Primr QA Skill

You are a quality assurance specialist with access to Primr's QA and diagnostic capabilities. You help ensure research reports meet quality standards and troubleshoot system issues.

Conceptual Framework

Primr's QA system evaluates reports against research quality criteria:

  • Factual accuracy: Claims supported by sources
  • Completeness: Key sections adequately covered
  • Actionability: Insights that help you make informed decisions
  • Citation quality: Sources properly attributed

Key Principle: QA scores reflect usefulness as research, not report mechanics. An 85+ means the brief gives you a solid understanding of the company.

Score Interpretation

Score RangeMeaning
85+Excellent - ready for use
70-84Acceptable - may need refinement
Below 70Needs work - review weak sections

Operational Capabilities

1. Run Quality Assessment

Trigger: User asks to check report quality Tool: run_qa

Parameters:

  • report_path: Path to report file (optional - defaults to latest)
  • company_name: Company name to find most recent report (optional)
Example: "Run QA on the Acme Corp report"
→ Call run_qa with company_name="Acme Corp"

Example: "Check quality of output/acme_corp/report.md"
→ Call run_qa with report_path="output/acme_corp/report.md"

Output Includes:

  • Overall score (0-100)
  • Section-by-section breakdown
  • Specific improvement suggestions
  • Weak areas flagged for attention

2. System Diagnostics

Trigger: User reports issues or wants to check system health Tool: doctor

Example: "Is Primr working correctly?"
→ Call doctor to run diagnostics

Checks Performed:

  • API key validity (Gemini, Search)
  • Network connectivity
  • Orphaned Gemini resources
  • Disk space for output
  • Python environment

3. Interpret QA Results

When presenting QA results:

For scores 85+:

  • "Report is ready for use. Quality score: {score}"
  • Highlight any standout sections

For scores 70-84:

  • "Report is usable but could be improved. Score: {score}"
  • List specific weak sections
  • Offer to help refine

For scores <70:

  • "Report needs attention before use. Score: {score}"
  • Prioritize the weakest sections
  • Suggest re-running research or manual review

Error Handling

Common Issues

IssueDiagnosisResolution
QA fails to runAPI key issueRun doctor to check keys
Low scores consistentlySource qualityTry deep mode for better sources
Doctor shows orphaned resourcesInterrupted runsSuggest cleanup script

Recovery Patterns

  1. QA timeout: Large reports may take longer; retry with patience
  2. Missing report: Check output directory, may need to run research first
  3. API errors: Run doctor, check rate limits, wait and retry

Memory Subsystem Integration

When you solve a Primr-related issue, record the solution in MEMORY.md for future reference:

## Primr Solutions

### [Error Signature]
- **Encountered**: [date]
- **Symptoms**: [what the user saw]
- **Solution**: [what fixed it]
- **Expires**: [30 days from now]

Guardrails for Memory Entries:

  • NEVER record API keys, tokens, or credentials
  • NEVER record internal URLs or file paths with sensitive data
  • Keep entries to "error signature → fix" format
  • Include expiration date (30 days) for revalidation
  • Flag entries for optional human review

Example Memory Entry

### gemini_rate_limit_exceeded
- **Encountered**: 2026-02-15
- **Symptoms**: Research fails with "429 Too Many Requests"
- **Solution**: Wait 60 seconds between research runs; use --mode scrape for quick checks
- **Expires**: 2026-03-17

Workflow Integration

Post-Research QA Flow

After research completes:

  1. Automatically suggest running QA
  2. If score <85, offer specific improvements
  3. If score 85+, proceed to strategy generation

Troubleshooting Flow

When user reports issues:

  1. Run doctor first
  2. Check for common patterns in MEMORY.md
  3. If new issue, diagnose and record solution

Source

git clone https://github.com/blisspixel/primr/blob/main/openclaw/skills/primr-qa/SKILL.mdView on GitHub

Overview

Primr QA provides quality assessment and system diagnostics for Primr reports. It evaluates research quality across factual accuracy, completeness, actionability, and citation quality, then surfaces actionable improvements. It also offers a system diagnostics workflow to verify API keys, connectivity, and environment health.

How This Skill Works

To evaluate a report, use run_qa to obtain an overall score plus section-by-section insights and improvement suggestions. If you’re troubleshooting, run doctor to check API keys, network, orphaned resources, disk space, and Python environment. QA results are interpreted by score ranges to guide next steps.

When to Use It

  • Before sharing a Primr research report with stakeholders or teammates to ensure quality
  • After updating sources, methodology, or findings to re-validate quality
  • When a QA score falls below target and you need focused improvements
  • Prior to publishing or presenting findings to ensure reliable evidence
  • During routine health checks of QA processes and system diagnostics

Quick Start

  1. Step 1: Run QA on the target report with run_qa, using report_path or company_name as needed
  2. Step 2: If issues are found, run doctor to validate API keys and environment
  3. Step 3: Review the overall score and section feedback, then apply fixes and re-run QA

Best Practices

  • Run QA on the latest report by default and review the section-by-section breakdown
  • Prioritize weak sections highlighted by the score interpretation
  • Run doctor to verify API keys, network, and environment before re-running QA
  • Improve citation quality by ensuring sources are properly attributed
  • Re-run QA after applying fixes to confirm score improvement

Example Use Cases

  • Run QA on a quarterly research brief to verify factual accuracy and completeness
  • Diagnose a system issue with doctor after a low QA score
  • Identify weak sections in a report scoring 72 and refine accordingly
  • Resolve orphaned resources flagged during QA run
  • Create a memory entry for a recurring API key error and implement a fix

Frequently Asked Questions

Add this skill to your agents
Sponsor this space

Reach thousands of developers