skill-upgrader
npx machina-cli add skill parcadei/Continuous-Claude-v3/skill-upgrader --openclawSkill Upgrader
Meta-skill that upgrades any SKILL.md to Decision Theory v5 Hybrid format using 4 parallel Ragie-backed agents.
When to Use
- "Upgrade this skill to v5"
- "Formalize this skill with decision theory"
- "Add MDP structure to this skill"
- "Apply the skill-upgrader to X"
Prerequisites
Ragie RAG with indexed books:
- decision-theory partition: LaValle Planning Algorithms, Sutton & Barto RL
- modal-logic partition: Blackburn Modal Logic, Huth & Ryan Logic in CS
Workflow
Step 1: Setup Session
SESSION=$(date +%Y%m%d-%H%M%S)-upgrade-{skill_name}
mkdir -p thoughts/skill-builds/${SESSION}
Step 2: Initialize Blackboard
Create thoughts/skill-builds/{session}/00-blackboard.md:
# Skill Upgrade: {skill_name}
Started: {timestamp}
## Input Skill
{path_to_skill}
## Target Format
Decision Theory v5 Hybrid
## Agent Findings
(Agents append below)
---
Step 3: Launch 4 Agents in Parallel
Use Task tool to spawn all 4 agents simultaneously. Each agent:
- Reads the input skill
- Queries Ragie for their specific book
- Appends findings to the blackboard
Agent 1: LaValle Planner
Book: LaValle's "Planning Algorithms" (decision-theory partition) Focus: States, Actions, Transitions
Task(
subagent_type="general-purpose",
prompt="""
INPUT SKILL: {path}
BLACKBOARD: thoughts/skill-builds/{session}/00-blackboard.md
YOUR BOOK: LaValle's "Planning Algorithms" in Ragie partition 'decision-theory'
TASK: Identify MDP structure in the skill.
Query Ragie:
```bash
uv run python scripts/ragie_query.py -q "MDP state space definition" -p decision-theory
uv run python scripts/ragie_query.py -q "action space sequential decisions" -p decision-theory
uv run python scripts/ragie_query.py -q "POMDP partial observability" -p decision-theory
Read the input skill and answer:
- What are the STATES? (phases, modes, tracked info)
- What are the ACTIONS? (what can agent do in each state)
- How do TRANSITIONS work? (deterministic or stochastic)
- Is this POMDP or fully observable?
WRITE to blackboard section: ## Agent 1: States, Actions & Transitions
Format as plain English with LaValle chapter citations. """ )
---
## Agent 2: Sutton & Barto Optimizer
**Book:** Sutton & Barto's "Reinforcement Learning" (decision-theory partition)
**Focus:** Policy, Termination, Value
**Depends on:** Agent 1
Task( subagent_type="general-purpose", prompt=""" INPUT SKILL: {path} BLACKBOARD: thoughts/skill-builds/{session}/00-blackboard.md
YOUR BOOK: Sutton & Barto's "Reinforcement Learning" in Ragie partition 'decision-theory'
WAIT: Read Agent 1's findings from blackboard first.
TASK: Design policy and termination conditions.
Query Ragie:
uv run python scripts/ragie_query.py -q "policy deterministic stochastic" -p decision-theory
uv run python scripts/ragie_query.py -q "episodic termination conditions" -p decision-theory
uv run python scripts/ragie_query.py -q "reward function design" -p decision-theory
Using Agent 1's states and actions, answer:
- What's the POLICY? (state → action rules)
- When does it END? (terminal states, success/failure)
- What are REWARDS? (goals +, costs -)
- Which states are HIGH/LOW value?
WRITE to blackboard section: ## Agent 2: Policy & Values
Format as plain English with Sutton & Barto section citations. """ )
---
## Agent 3: Blackburn Modal Logician
**Book:** Blackburn's "Modal Logic" (modal-logic partition)
**Focus:** Constraints (temporal, epistemic, deontic)
Task( subagent_type="general-purpose", prompt=""" INPUT SKILL: {path} BLACKBOARD: thoughts/skill-builds/{session}/00-blackboard.md
YOUR BOOK: Blackburn's "Modal Logic" in Ragie partition 'modal-logic'
TASK: Extract constraints from the skill.
Query Ragie:
uv run python scripts/ragie_query.py -q "temporal logic LTL operators" -p modal-logic
uv run python scripts/ragie_query.py -q "epistemic logic knowledge" -p modal-logic
uv run python scripts/ragie_query.py -q "deontic logic obligations" -p modal-logic
Read the input skill and identify:
- TEMPORAL: "must do X before Y" → □, ◇, U
- EPISTEMIC: "must know X" → K operator
- DEONTIC: "must/forbidden/may" → O, F, P
- DYNAMIC: "action causes effect" → [action]
WRITE to blackboard section: ## Agent 3: Constraints
For each constraint:
- Plain English description
- Modal logic notation
- Why it matters
- Blackburn chapter citation """ )
---
## Agent 4: Huth & Ryan Verifier
**Book:** Huth & Ryan's "Logic in Computer Science" (modal-logic partition)
**Focus:** Validation, Safety, Liveness
**Depends on:** Agents 1-3
Task( subagent_type="general-purpose", prompt=""" INPUT SKILL: {path} BLACKBOARD: thoughts/skill-builds/{session}/00-blackboard.md
YOUR BOOK: Huth & Ryan's "Logic in Computer Science" in Ragie partition 'modal-logic'
WAIT: Read Agents 1-3 findings from blackboard first.
TASK: Verify consistency and completeness.
Query Ragie:
uv run python scripts/ragie_query.py -q "safety properties verification" -p modal-logic
uv run python scripts/ragie_query.py -q "liveness properties eventually" -p modal-logic
uv run python scripts/ragie_query.py -q "model checking CTL" -p modal-logic
Check:
- SAFETY: What bad things never happen? □¬(bad)
- LIVENESS: What good things eventually happen? ◇(good)
- CONSISTENCY: Any contradictions between agents?
- COMPLETENESS: Any gaps in coverage?
WRITE to blackboard section: ## Agent 4: Verification
Report with ✓/✗ for each property. Overall verdict: PASS or NEEDS_WORK Huth & Ryan section citations. """ )
---
## Step 4: Synthesize Final Skill
After all agents complete, read the blackboard and create:
**Output:** `thoughts/skill-builds/{session}/SKILL-upgraded.md`
Use v5 Hybrid template:
```yaml
---
name: {original_name}
description: {original_description}
version: 5.1-hybrid
---
# Option: {name}
## Initiation (I)
[From original + Agent 1 state analysis]
## Observation Space (Y)
[From Agent 1 POMDP analysis]
## Action Space (U)
[From Agent 1 actions]
## Policy (pi)
[From Agent 2 state→action rules]
## Termination (beta)
[From Agent 2 episode structure]
## Q-Heuristics
[From Agent 2 value guidance]
## Constraints
[From Agent 3 modal logic]
## Verification
[From Agent 4 safety/liveness]
Example Usage
User: "Upgrade .claude/skills/implement_plan/SKILL.md to v5 Hybrid"
Claude:
1. Creates session directory
2. Initializes blackboard
3. Launches 4 agents in parallel (Task tool)
4. Waits for completion
5. Reads blackboard
6. Synthesizes upgraded skill
7. Reports: "Upgraded skill at thoughts/skill-builds/.../SKILL-upgraded.md"
Ragie Query Reference
# Decision theory partition
uv run python scripts/ragie_query.py -q "your question" -p decision-theory
# Modal logic partition
uv run python scripts/ragie_query.py -q "your question" -p modal-logic
# With reranking for better results
uv run python scripts/ragie_query.py -q "your question" -p decision-theory --rerank
Files Created
After upgrade:
thoughts/skill-builds/{session}/
├── 00-blackboard.md # Agent collaboration
├── SKILL-upgraded.md # Final v5 Hybrid skill
└── validation-report.md # Agent 4 verification
Source
git clone https://github.com/parcadei/Continuous-Claude-v3/blob/main/.claude/skills/skill-upgrader/SKILL.mdView on GitHub Overview
Skill Upgrader is a meta-skill that transforms any SKILL.md into a Decision Theory v5 Hybrid format. It coordinates four Ragie-backed agents to extract states, actions, transitions, policies, rewards, and constraints, then records findings on a shared blackboard to produce a fully structured upgrade.
How This Skill Works
The process sets up a session and a blackboard, then launches four Ragie-backed agents in parallel. Each agent reads the input skill, queries Ragie books for relevant theory, and appends findings to the blackboard. The agents cover MDp structure, policy/value design, and modal-logic constraints to compose the v5 Hybrid format.
When to Use It
- Upgrade this skill to v5
- Formalize this skill with decision theory
- Add MDP structure to this skill
- Apply the skill-upgrader to X
- Convert a non-format skill into a Decision Theory v5 Hybrid blueprint
Quick Start
- Step 1: Run the upgrader on your target SKILL.md
- Step 2: The 4 Ragie-backed agents populate thoughts/skill-builds/{SESSION}/00-blackboard.md
- Step 3: Review and export the upgraded SKILL.md in Decision Theory v5 Hybrid format
Best Practices
- Keep the source SKILL.md path clearly referenced in the session
- Initialize the blackboard with the skill name and target format
- Review Agent 1 (MDP) findings for consistency with the skill goals
- Cross-check Agent 2 outputs (policy, termination, rewards) against objectives
- Document citations (e.g., LaValle, Sutton & Barto, Blackburn) in the final SKILL.md
Example Use Cases
- Upgrade a simple 'pricing-skill' to a v5 Hybrid with explicit states, actions, and rewards
- Formalize a 'warehouse-routing-skill' by adding MDP structure and transition dynamics
- Add modal-logic constraints to an 'access-control-skill' for temporal and deontic rules
- Transform a 'data-cleaning-skill' into a decision-theory guided hybrid plan
- Convert a 'customer-support-skill' into a v5 Hybrid format with policy and value mappings