Get the FREE Ultimate OpenClaw Setup Guide →

metric-dashboard

Scanned
npx machina-cli add skill aroyburman-codes/pm-skills/metric-dashboard --openclaw
Files (1)
SKILL.md
4.5 KB

Metric Dashboard Skill

Design a comprehensive metric dashboard and KPI tracking plan for any product or feature.

When to Use

  • User needs to define metrics for a new product or feature
  • User is setting up monitoring and alerting
  • User needs to design a dashboard layout
  • User says /metric-dashboard followed by the product/feature
  • Any time measurement strategy needs to be defined

Framework: Metric Dashboard Design (5 Steps)

Step 1: Define the Metric Hierarchy

North Star Metric (NSM): The single metric that best captures the value your product delivers.

  • Must reflect user value, not just business value
  • Must be measurable with current instrumentation
  • Formula: NSM = [engagement unit] per [user segment] per [time period]

Decompose into a metric tree:

North Star Metric
├── Input Metric A (e.g., new users)
│   ├── Sub-metric A1
│   └── Sub-metric A2
├── Input Metric B (e.g., activation rate)
│   ├── Sub-metric B1
│   └── Sub-metric B2
└── Input Metric C (e.g., retention)
    ├── Sub-metric C1
    └── Sub-metric C2

Step 2: Categorize Metrics

Product Metrics:

  • Acquisition: How users find you (sign-ups, installs, registrations)
  • Activation: First value moment (onboarding completion, first action)
  • Engagement: Core usage (DAU/MAU, session length, feature adoption)
  • Retention: Coming back (D1/D7/D30, cohort retention curves)
  • Revenue: Monetization (ARPU, conversion, LTV, churn)

Technical Metrics:

  • Performance: Latency (p50, p95, p99), throughput, error rate
  • Reliability: Uptime, incident count, MTTR
  • Infrastructure: CPU/memory utilization, cost per request

AI/ML Metrics (if applicable):

  • Quality: Accuracy, hallucination rate, eval scores
  • Safety: Content policy violation rate, false refusal rate
  • Cost: Cost per inference, token usage
  • Latency: Time to first token, tokens per second

Business Metrics:

  • Revenue: MRR, ARR, revenue growth rate
  • Unit economics: CAC, LTV, LTV/CAC ratio
  • Market: Market share, competitive win rate

Step 3: Set Targets & Alerts

For each metric, define:

MetricCurrentTargetAlert ThresholdOwner
NSMXYZPM
Metric A
Metric B

Alert levels:

  • Warning (yellow): Metric trending below target — investigate
  • Critical (red): Metric below threshold — immediate action required
  • Anomaly: Unexpected spike or drop — auto-detect and notify

Step 4: Design Dashboard Layout

Executive Dashboard (1 screen):

  • NSM trend (last 30/90 days) — large, prominent
  • 4-6 key metrics with sparklines and trend arrows
  • Traffic light status (green/yellow/red) for each area
  • Notable events annotated on the timeline

Operational Dashboard (detailed):

  • Real-time metrics for the current day/hour
  • Breakdowns by segment (platform, geography, user type)
  • Funnel visualization (acquisition → activation → retention)
  • Experiment results (A/B test outcomes)

Technical Dashboard (if applicable):

  • System health (latency, error rate, uptime)
  • Model performance (eval scores, cost, throughput)
  • Infrastructure utilization and cost

Step 5: Measurement Plan

For each metric, document:

  • Definition: Exact formula, including/excluding criteria
  • Data source: Which event, table, or API
  • Instrumentation: What needs to be logged/tracked
  • Granularity: How often updated (real-time, hourly, daily)
  • Segments: Key breakdowns (platform, country, user tier)
  • Owner: Who monitors this metric

Output Format

Generate a complete metric plan in markdown with:

  1. Metric hierarchy (tree diagram)
  2. Metric definitions table
  3. Targets and alert thresholds
  4. Dashboard layout description
  5. Measurement plan

Common Pitfalls to Avoid

  • Vanity metrics: Big numbers that don't reflect value (total sign-ups vs. active users)
  • Too many metrics: 5-8 key metrics max on the exec dashboard
  • No baselines: Always show current state before setting targets
  • Missing guardrails: Every optimization metric needs a counter-metric
  • No segmentation: Averages hide problems — always break down by segment

Source

git clone https://github.com/aroyburman-codes/pm-skills/blob/main/skills/metric-dashboard/SKILL.mdView on GitHub

Overview

Design a comprehensive metric dashboard and KPI tracking plan for any product or feature. It defines what to measure, how to measure it, alert thresholds, and dashboard layout across product, business, and technical metrics.

How This Skill Works

Begin by defining the North Star Metric (NSM) and decomposing it into a metric tree. Next, categorize metrics into Product, Technical, AI/ML, and Business, set targets and alert thresholds, and design executive and operational dashboards. Finally, document a full measurement plan detailing definitions, data sources, instrumentation, granularity, segments, and ownership.

When to Use It

  • User needs to define metrics for a new product or feature
  • User is setting up monitoring and alerting
  • User needs to design a dashboard layout
  • User says "/metric-dashboard" followed by the product/feature
  • Any time measurement strategy needs to be defined

Quick Start

  1. Step 1: Define the North Star Metric and build the metric tree with inputs and sub-metrics
  2. Step 2: Categorize metrics (Product, Technical, AI/ML, Business), set targets, and establish alert levels
  3. Step 3: Design executive and operational dashboards and finalize the measurement plan for each metric

Best Practices

  • Define a North Star Metric that truly reflects user value and is measurable with existing instrumentation
  • Decompose the NSM into a metric tree (inputs and sub-metrics) to map cause-and-effect
  • Categorize metrics by Product, Technical, AI/ML, and Business, and assign clear owners
  • Set explicit targets and multi-level alerts (Warning, Critical, Anomaly) for every metric
  • Document a full measurement plan for each metric: definition, data source, instrumentation, granularity, segments, and ownership

Example Use Cases

  • SaaS onboarding: NSM as engagement per user per month; track activation, retention, and revenue metrics with alerts on drift
  • Mobile app: NSM as sessions per user per day; monitor DAU/MAU, session length, feature adoption, and retention curves
  • AI product: NSM as successful inferences per user per hour; monitor quality (accuracy), latency, cost per inference, and safety
  • E-commerce: NSM as purchases per user per period; track conversion, ARPU, LTV, and churn across segments
  • Platform infrastructure: NSM as requests per second per user; monitor latency (p50/p95/p99), error rate, uptime, and cost

Frequently Asked Questions

Add this skill to your agents
Sponsor this space

Reach thousands of developers