Get the FREE Ultimate OpenClaw Setup Guide →

ln-230-story-prioritizer

Scanned
npx machina-cli add skill levnikolaevich/claude-code-skills/ln-230-story-prioritizer --openclaw
Files (1)
SKILL.md
12.5 KB

Paths: File paths (shared/, references/, ../ln-*) are relative to skills repo root. If not found at CWD, locate this SKILL.md directory and go up one level for repo root.

Story Prioritizer

Evaluate Stories using RICE scoring with market research. Generate consolidated prioritization table for Epic.

Purpose & Scope

  • Prioritize Stories AFTER ln-220 creates them
  • Research market size and competition per Story
  • Calculate RICE score for each Story
  • Generate prioritization table (P0/P1/P2/P3)
  • Output: docs/market/[epic-slug]/prioritization.md

When to Use

Use this skill when:

  • Stories created by ln-220, need business prioritization
  • Planning sprint with limited capacity (which Stories first?)
  • Stakeholder review requires data-driven priorities
  • Evaluating feature ROI before implementation

Do NOT use when:

  • Epic has no Stories yet (run ln-220 first)
  • Stories are purely technical (infrastructure, refactoring)
  • Prioritization already exists in docs/market/

Who calls this skill:

  • ln-200-scope-decomposer Phase 4 (optional, sequential per Epic)
  • User (manual) - standalone after ln-220-story-coordinator

Input Parameters

ParameterRequiredDescriptionDefault
epicYesEpic ID or "Epic N" format-
storiesNoSpecific Story IDs to prioritizeAll in Epic
depthNoResearch depth (quick/standard/deep)"standard"

depth options:

  • quick - 2-3 min/Story, 1 WebSearch per type
  • standard - 5-7 min/Story, 2-3 WebSearches per type
  • deep - 8-10 min/Story, comprehensive research

Output Structure

docs/market/[epic-slug]/
└── prioritization.md    # Consolidated table + RICE details + sources

Table columns (from user requirements):

PriorityCustomer ProblemFeatureSolutionRationaleImpactMarketSourcesCompetition
P0User pain pointStory titleTechnical approachWhy importantBusiness impact$XB[Link]Blue 1-3 / Red 4-5

Research Tools

ToolPurposeExample Query
WebSearchMarket size, competitors"[domain] market size {current_year}"
mcp__RefIndustry reports"[domain] market analysis report"
LinearLoad Storieslist_issues(project=Epic.id)
GlobCheck existing"docs/market/[epic]/*"

Workflow

Phase 1: Discovery (2 min)

Objective: Validate input and prepare context.

Process:

  1. Parse Epic input:

    • Accept: Epic ID, "Epic N", or Linear Project URL
    • Query: get_project(query=epic)
    • Extract: Epic ID, title, description
  2. Auto-discover configuration:

    • Read docs/tasks/kanban_board.md for Team ID
    • Slugify Epic title for output path
  3. Check existing prioritization:

    Glob: docs/market/[epic-slug]/prioritization.md
    
    • If exists: Ask "Update existing or create new?"
    • If new: Continue
  4. Create output directory:

    mkdir -p docs/market/[epic-slug]/
    

Output: Epic metadata, output path, existing check result


Phase 2: Load Stories Metadata (3 min)

Objective: Build Story queue with metadata only (token efficiency).

Process:

  1. Query Stories from Epic:

    list_issues(project=Epic.id, label="user-story")
    
  2. Extract metadata only:

    • Story ID, title, status
    • DO NOT load full descriptions yet
  3. Filter Stories:

    • Exclude: Done, Cancelled, Archived
    • Include: Backlog, Todo, In Progress
  4. Build processing queue:

    • Order by: existing priority (if any), then by ID
    • Count: N Stories to process

Output: Story queue (ID + title), ~50 tokens/Story


Phase 3: Story-by-Story Analysis Loop (5-10 min/Story)

Objective: For EACH Story: load description, research, score RICE.

Critical: Process Stories ONE BY ONE for token efficiency!

Per-Story Steps:

Step 3.1: Load Story Description
get_issue(id=storyId, includeRelations=false)

Extract from Story:

  • Feature: Story title
  • Customer Problem: From "So that [value]" + Context section
  • Solution: From Technical Notes (implementation approach)
  • Rationale: From AC + Success Criteria
Step 3.2: Research Market Size

WebSearch queries (based on depth):

"[customer problem domain] market size TAM {current_year}"
"[feature type] industry market forecast"

mcp__Ref query:

"[domain] market analysis Gartner Statista"

Extract:

  • Market size: $XB (with unit: B=Billion, M=Million)
  • Growth rate: X% CAGR
  • Sources: URL + date

Confidence mapping:

  • Industry report (Gartner, Statista) → Confidence 0.9-1.0
  • News article → Confidence 0.7-0.8
  • Blog/Forum → Confidence 0.5-0.6
Step 3.3: Research Competition

WebSearch queries:

"[feature] competitors alternatives {current_year}"
"[solution approach] market leaders"

Count competitors and classify:

Competitors FoundCompetition IndexOcean Type
01Blue Ocean
1-22Emerging
3-53Growing
6-104Mature
>105Red Ocean
Step 3.4: Calculate RICE Score
RICE = (Reach x Impact x Confidence) / Effort

Reach (1-10): Users affected per quarter

ScoreUsersIndicators
1-2<500Niche, single persona
3-4500-2KDepartment-level
5-62K-5KOrganization-wide
7-85K-10KMulti-org
9-10>10KPlatform-wide

Impact (0.25-3.0): Business value

ScoreLevelIndicators
0.25MinimalNice-to-have
0.5LowQoL improvement
1.0MediumEfficiency gain
2.0HighRevenue driver
3.0MassiveStrategic differentiator

Confidence (0.5-1.0): Data quality (from Step 3.2)

Data Confidence Assessment:

For each RICE factor, assess data confidence level:

ConfidenceCriteriaScore Modifier
HIGHMultiple authoritative sources (Gartner, Statista, SEC filings)Factor used as-is
MEDIUM1-2 sources, mixed quality (blog + report)Factor ±25% range shown
LOWNo sources, team estimate onlyFactor ±50% range shown

Output: Show confidence per factor in prioritization table + RICE range (optimistic/pessimistic) to make uncertainty explicit.

Effort (1-10): Person-months

ScoreTimeStory Indicators
1-2<2 weeks3 AC, simple CRUD
3-42-4 weeks4 AC, integration
5-61-2 months5 AC, complex logic
7-82-3 monthsExternal dependencies
9-103+ monthsNew infrastructure
Step 3.5: Determine Priority
PriorityRICE ThresholdCompetition Override
P0 (Critical)>= 30OR Competition = 1 (Blue Ocean monopoly)
P1 (High)>= 15OR Competition <= 2 (Emerging market)
P2 (Medium)>= 5-
P3 (Low)< 5Competition = 5 (Red Ocean) forces P3
Step 3.6: Store and Clear
  • Append row to in-memory results table
  • Clear Story description from context
  • Move to next Story in queue

Output per Story: Complete row for prioritization table


Phase 4: Generate Prioritization Table (5 min)

Objective: Create consolidated markdown output.

Process:

  1. Sort results:

    • Primary: Priority (P0 → P3)
    • Secondary: RICE score (descending)
  2. Generate markdown:

    • Use template from references/prioritization_template.md
    • Fill: Priority Summary, Main Table, RICE Details, Sources
  3. Save file:

    Write: docs/market/[epic-slug]/prioritization.md
    

Output: Saved prioritization.md


Phase 5: Summary & Next Steps (1 min)

Objective: Display results and recommendations.

Output format:

## Prioritization Complete

**Epic:** [Epic N - Name]
**Stories analyzed:** X
**Time elapsed:** Y minutes

### Priority Distribution:
- P0 (Critical): X Stories - Implement ASAP
- P1 (High): X Stories - Next sprint
- P2 (Medium): X Stories - Backlog
- P3 (Low): X Stories - Consider deferring

### Top 3 Priorities:
1. [Story Title] - RICE: X, Market: $XB, Competition: Blue/Red

### Saved to:
docs/market/[epic-slug]/prioritization.md

### Next Steps:
1. Review table with stakeholders
2. Run ln-300 for P0/P1 Stories first
3. Consider cutting P3 Stories

Time-Box Constraints

DepthPer-StoryTotal (10 Stories)
quick2-3 min20-30 min
standard5-7 min50-70 min
deep8-10 min80-100 min

Time management rules:

  • If Story exceeds time budget: Skip deep research, use estimates (Confidence 0.5)
  • If total exceeds budget: Switch to "quick" depth for remaining Stories
  • Parallel WebSearch where possible (market + competition)

Token Efficiency

Loading pattern:

  • Phase 2: Metadata only (~50 tokens/Story)
  • Phase 3: Full description ONE BY ONE (~3,000-5,000 tokens/Story)
  • After each Story: Clear description, keep only result row (~100 tokens)

Memory management:

  • Sequential processing (not parallel)
  • Maximum context: 1 Story description at a time
  • Results accumulate as compact table rows

Integration with Ecosystem

Position in workflow:

ln-210 (Scope → Epics)
     ↓
ln-220 (Epic → Stories)
     ↓
ln-230 (RICE per Story → prioritization table) ← THIS SKILL
     ↓
ln-300 (Story → Tasks)

Dependencies:

  • WebSearch, mcp__Ref (market research)
  • Linear MCP (load Epic, Stories)
  • Glob, Write, Bash (file operations)

Downstream usage:

  • Sprint planning uses P0/P1 to select Stories
  • ln-300 processes Stories in priority order
  • Stakeholders review before implementation

Critical Rules

  1. Source all data - Every Market number needs source + date
  2. Prefer recent data - last 2 years, warn if older
  3. Cross-reference - 2+ sources for Market size (reduce error)
  4. Time-box strictly - Skip depth for speed if needed
  5. Confidence levels - Mark High/Medium/Low for estimates
  6. No speculation - Only sourced claims, note "[No data]" gaps
  7. One Story at a time - Token efficiency critical
  8. Preserve language - If user asks in Russian, respond in Russian

Definition of Done

  • Epic validated in Linear
  • All Stories loaded (metadata, then descriptions per-Story)
  • Market research completed (2+ sources per Story)
  • RICE score calculated for each Story
  • Competition index assigned (1-5)
  • Priority assigned (P0/P1/P2/P3)
  • Table sorted by Priority + RICE
  • File saved to docs/market/[epic-slug]/prioritization.md
  • Summary with top priorities and next steps
  • Total time within budget

Example Usage

Basic usage:

ln-230-story-prioritizer epic="Epic 7"

With parameters:

ln-230-story-prioritizer epic="Epic 7: Translation API" depth="deep"

Specific Stories:

ln-230-story-prioritizer epic="Epic 7" stories="US001,US002,US003"

Example output (docs/market/translation-api/prioritization.md):

PriorityCustomer ProblemFeatureSolutionRationaleImpactMarketSourcesCompetition
P0"Repeat translations cost GPU"Translation MemoryRedis cache, 5ms lookup70-90% GPU cost reductionHigh$2B+M&M3
P0"Can't translate PDF"PDF SupportPDF parsing + layoutEnterprise blockerHigh$10B+Eden5
P1"Need video subtitles"SRT/VTT SupportTiming preservationBlue Ocean opportunityMedium$5.7BGMI2

Reference Files

FilePurpose
prioritization_template.mdOutput markdown template
rice_scoring_guide.mdRICE factor scales and examples
research_queries.mdWebSearch query templates by domain
competition_index.mdBlue/Red Ocean classification rules

Version: 1.0.0 Last Updated: 2025-12-23

Source

git clone https://github.com/levnikolaevich/claude-code-skills/blob/master/ln-230-story-prioritizer/SKILL.mdView on GitHub

Overview

Evaluates each user story using RICE scoring and integrates market research to generate a consolidated prioritization table for an Epic. The output is written to docs/market/[epic-slug]/prioritization.md and the workflow runs after ln-220's story creation.

How This Skill Works

Triggered as an L2 worker after ln-220, the skill discovers the Epic, output path, and existing prioritization, then loads story metadata, performs market research, computes RICE scores for each story, and generates a consolidated prioritization table (P0–P3) with sources and competitive context stored in docs/market/[epic-slug]/prioritization.md.

When to Use It

  • After ln-220 creates Stories, to assign business prioritization for the Epic
  • Sprint planning with limited capacity to decide which stories ship first
  • Stakeholder reviews requiring data-driven prioritization
  • Evaluating feature ROI before committing resources
  • Regenerate prioritization when market signals or priorities shift

Quick Start

  1. Step 1: Provide the epic identifier and optional depth, then trigger ln-230-story-prioritizer after ln-220 creates stories
  2. Step 2: The skill loads story metadata, runs market research, and computes RICE scores for each story
  3. Step 3: Open docs/market/[epic-slug]/prioritization.md to review and adjust the consolidated table

Best Practices

  • Ensure the epic slug and output path are correct before running
  • Choose depth quick, standard, or deep based on expected research effort per story
  • Include all relevant stories in the backlog and exclude done or cancelled items
  • Keep the prioritization sources and competition data up to date in prioritization.md
  • Validate alignment with business goals before finalizing the plan

Example Use Cases

  • Epic on onboarding: P0 reduces signup drop-off, P1 improves guided tour, P2 adds email onboarding, P3 enhances micro-interactions
  • Epic on analytics dashboard: P0 real-time charts, P1 customizable reports, P2 data export, P3 widget theming
  • Epic on mobile payments: P0 frictionless checkout, P1 receipt automation, P2 offline mode, P3 multi-currency support
  • Epic on chat automation: P0 proactive chat prompts, P1 knowledge base integration, P2 sentiment analysis, P3 agent handoff
  • Epic on inventory alerts: P0 real-time stock alerts, P1 demand forecasting, P2 auto-reorder, P3 supplier portal links

Frequently Asked Questions

Add this skill to your agents
Sponsor this space

Reach thousands of developers