feature-prioritization

Prioritization Framework Application

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "feature-prioritization" with this command: npx skills add rsmdt/the-startup/rsmdt-the-startup-feature-prioritization

Prioritization Framework Application

Systematic frameworks for making objective prioritization decisions that balance value, effort, and strategic alignment.

When to Activate

  • Prioritizing feature backlogs

  • Evaluating competing initiatives

  • Making build vs defer decisions

  • Creating product roadmaps

  • Allocating limited resources

  • Justifying prioritization decisions to stakeholders

RICE Framework

Quantitative scoring for comparing initiatives objectively.

Formula

RICE Score = (Reach × Impact × Confidence) / Effort

Components

Factor Description Scale

Reach How many users affected per quarter Actual number (100, 1000, 10000)

Impact Effect on each user 0.25 (Minimal) to 3 (Massive)

Confidence How sure are we 50% (Low) to 100% (High)

Effort Person-months required Actual estimate (0.5, 1, 3, 6)

Impact Scale

Score Label Description

3 Massive Life-changing for users, core workflow transformation

2 High Major improvement, significant time savings

1 Medium Noticeable improvement, minor friction reduction

0.5 Low Slight improvement, nice-to-have

0.25 Minimal Barely noticeable difference

Confidence Scale

Score Label Basis

100% High User research + validated data + successful tests

80% Medium Some data + team experience + analogous examples

50% Low Intuition only, no supporting data

Example Calculation

Feature: One-click reorder

Reach: 5,000 (customers who reorder monthly) Impact: 2 (High - saves significant time) Confidence: 80% (Based on support ticket analysis) Effort: 1 person-month

RICE = (5000 × 2 × 0.8) / 1 = 8000

Feature: Dark mode

Reach: 20,000 (all active users) Impact: 0.5 (Low - preference, not productivity) Confidence: 50% (No data, user requests only) Effort: 2 person-months

RICE = (20000 × 0.5 × 0.5) / 2 = 2500

Decision: One-click reorder scores higher, prioritize first

RICE Template

Feature Reach Impact Confidence Effort Score Rank

Feature A 5000 2 80% 1 8000 1

Feature B 20000 0.5 50% 2 2500 2

Value vs Effort Matrix

Visual framework for quick categorization.

The Matrix

          High Value
               │
┌──────────────┼──────────────┐
│              │              │
│  QUICK WINS  │  STRATEGIC   │
│  Do First    │  Plan & Do   │
│              │              │
├──────────────┼──────────────┤ High

Low │ │ │ Effort Effort │ │ │ FILL-INS │ TIME SINKS │ │ If Spare │ Avoid │ │ Capacity │ │ │ │ │ └──────────────┼──────────────┘ │ Low Value

Quadrant Actions

Quadrant Characteristics Action

Quick Wins High value, low effort Do immediately

Strategic High value, high effort Plan carefully, staff appropriately

Fill-Ins Low value, low effort Do when nothing else is ready

Time Sinks Low value, high effort Don't do (or simplify drastically)

Estimation Guidance

Value Assessment:

  • Revenue impact

  • Cost reduction

  • User satisfaction improvement

  • Strategic alignment

  • Risk reduction

Effort Assessment:

  • Development time

  • Design complexity

  • Testing requirements

  • Deployment complexity

  • Ongoing maintenance

Kano Model

Categorize features by their impact on satisfaction.

Categories

Satisfaction ▲ │ ╱ Delighters │ ╱ (Unexpected features) │ ╱ ─────┼────●──────────────────────────► Feature │ │╲ Implementation │ │ ╲ Performance │ │ (More is better) │ │ │ └── Must-Haves │ (Expected, dissatisfaction if missing) ▼

Category Definitions

Category Present Absent Example

Must-Have Neutral Very dissatisfied Login functionality

Performance More = better Less = worse Page load speed

Delighter Very satisfied Neutral Personalized recommendations

Indifferent No effect No effect Backend tech choice

Reverse Dissatisfied Satisfied Forced tutorials

Kano Survey Questions

For each feature, ask two questions:

Functional: "If [feature] were present, how would you feel?" Dysfunctional: "If [feature] were absent, how would you feel?"

Answer Options:

  1. I like it
  2. I expect it
  3. I'm neutral
  4. I can tolerate it
  5. I dislike it

Interpretation Matrix

Like Expect Neutral Tolerate Dislike

Like Q A A A O

Expect R I I I M

Neutral R I I I M

Tolerate R I I I M

Dislike R R R R Q

Key: M=Must-Have, O=One-dimensional, A=Attractive, I=Indifferent, R=Reverse, Q=Questionable

MoSCoW Method

Simple categorization for scope definition.

Categories

Category Definition Negotiability

Must Critical for success, release blocked without Non-negotiable

Should Important but not critical Can defer to next release

Could Nice to have, minor impact First to cut if needed

Won't Explicitly excluded from scope Not this release

Application Rules

Budget Allocation (Recommended):

  • Must: 60% of capacity
  • Should: 20% of capacity
  • Could: 20% of capacity (buffer)
  • Won't: 0% (explicitly excluded)

Why the buffer matters:

  • Must items often take longer than estimated
  • Should items may become Must if requirements change
  • Could items fill capacity at sprint end

Example

Feature: User Registration

MUST: ✓ Email/password signup ✓ Email verification ✓ Password requirements enforcement

SHOULD: ○ Social login (Google) ○ Remember me functionality ○ Password strength indicator

COULD: ◐ Social login (Facebook, Apple) ◐ Profile picture upload ◐ Username suggestions

WON'T (this release): ✗ Two-factor authentication ✗ SSO integration ✗ Biometric login

Cost of Delay

Prioritize by economic impact of waiting.

CD3 Formula

CD3 = Cost of Delay / Duration

Cost of Delay: Weekly value lost by not having the feature Duration: Weeks to implement

Delay Cost Types

Type Description Calculation

Revenue Sales not captured Lost deals × average value

Cost Ongoing expenses Weekly operational cost

Risk Penalty or loss potential Probability × impact

Opportunity Market window Revenue × time sensitivity

Urgency Profiles

            Value
              │

Standard: │──────────────── │ └──────────────────► Time

Urgent: │╲ │ ╲ │ ╲────────── │ └──────────────────► Time

Deadline: │ │────────┐ │ │ │ └─ (drops to zero) └──────────────────► Time

Example

Feature A: New payment method

  • Cost of Delay: $10,000/week (lost sales to competitor)
  • Duration: 4 weeks
  • CD3 = 10000 / 4 = 2500

Feature B: Admin dashboard redesign

  • Cost of Delay: $2,000/week (support inefficiency)
  • Duration: 2 weeks
  • CD3 = 2000 / 2 = 1000

Feature C: Compliance update (deadline in 6 weeks)

  • Cost of Delay: $50,000/week after deadline (fines)
  • Duration: 4 weeks
  • CD3 = 50000 / 4 = 12500 (if started now, 0 if after deadline)

Priority: C (deadline), then A (highest CD3), then B

Weighted Scoring

Custom scoring for organization-specific criteria.

Building a Weighted Model

Step 1: Define Criteria

  • Strategic alignment
  • Revenue potential
  • User demand
  • Technical feasibility
  • Competitive advantage

Step 2: Assign Weights (total = 100%)

CriterionWeight
Strategic30%
Revenue25%
User demand20%
Feasibility15%
Competitive10%

Step 3: Score Each Feature (1-5 scale)

FeatureStrategicRevenueDemandFeasibleCompetitiveTotal
A543423.95
B355333.90
C434543.85

Calculation

Score = Σ (criterion_score × criterion_weight)

Feature A: = (5 × 0.30) + (4 × 0.25) + (3 × 0.20) + (4 × 0.15) + (2 × 0.10) = 1.5 + 1.0 + 0.6 + 0.6 + 0.2 = 3.9

Decision Documentation

Priority Decision Record

Priority Decision: [Feature/Initiative]

Date: [YYYY-MM-DD]

Decision: [Prioritize / Defer / Reject]

Context

[What prompted this decision?]

Evaluation

Framework Used: [RICE / Kano / MoSCoW / Weighted]

Scores

[Show calculations or categorization]

Trade-offs Considered

  • Option A: [description] - [pros/cons]
  • Option B: [description] - [pros/cons]

Decision Rationale

[Why this priority over alternatives?]

Stakeholders

  • Agreed: [names]
  • Disagreed: [names, reasons documented]

Review Date

[When to revisit if deferred]

Framework Selection Guide

Situation Recommended Framework

Comparing many similar features RICE (quantitative)

Quick triage of backlog Value vs Effort

Understanding user expectations Kano Model

Defining release scope MoSCoW

Time-sensitive decisions Cost of Delay

Organization-specific criteria Weighted Scoring

Anti-Patterns

Anti-Pattern Problem Solution

HiPPO Highest-paid person's opinion wins Use data-driven frameworks

Recency Bias Last request gets priority Systematic evaluation of all options

Squeaky Wheel Loudest stakeholder wins Weight by strategic value

Analysis Paralysis Over-analyzing decisions Time-box evaluation

Sunken Cost Continuing failed initiatives Evaluate future value only

Feature Factory Shipping without measuring Tie features to outcomes

Best Practices

  • Use multiple frameworks - Validate with different approaches

  • Document decisions - Enable future learning

  • Revisit regularly - Priorities change as context evolves

  • Include stakeholders - Ensure buy-in

  • Measure outcomes - Validate prioritization quality

References

  • RICE Scoring Template - Spreadsheet template

  • Prioritization Workshop Guide - Facilitation guide

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

General

specify-requirements

No summary provided by upstream source.

Repository SourceNeeds Review
General

review

No summary provided by upstream source.

Repository SourceNeeds Review
General

analyze

No summary provided by upstream source.

Repository SourceNeeds Review