prioritization-frameworks

Prioritization Frameworks

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "prioritization-frameworks" with this command: npx skills add yonatangross/orchestkit/yonatangross-orchestkit-prioritization-frameworks

Prioritization Frameworks

Quantitative and qualitative frameworks for ranking features, initiatives, and backlog items.

RICE Framework

Developed by Intercom, RICE provides a data-driven score for comparing features.

Formula

RICE Score = (Reach × Impact × Confidence) / Effort

Factors

Factor Definition Scale

Reach Users/customers affected per quarter Actual number

Impact Effect on individual user 0.25 (minimal) to 3 (massive)

Confidence How sure are you? 0.5 (low) to 1.0 (high)

Effort Person-months required Actual estimate

Impact Scale

Score Level Description

3 Massive Fundamental improvement

2 High Significant improvement

1 Medium Noticeable improvement

0.5 Low Minor improvement

0.25 Minimal Barely noticeable

Confidence Scale

Score Level Evidence

1.0 High Strong data, validated

0.8 Medium Some data, reasonable assumptions

0.5 Low Gut feeling, little data

Example Calculation

Feature: Smart search with AI suggestions

Reach: 50,000 users/quarter (active searchers) Impact: 2 (high - significantly better results) Confidence: 0.8 (tested in prototype) Effort: 3 person-months

RICE = (50,000 × 2 × 0.8) / 3 = 26,667

RICE Template

FeatureReachImpactConfidenceEffortRICE Score
Feature A10,00020.828,000
Feature B50,00011.0412,500
Feature C5,00030.517,500

ICE Framework

Simpler than RICE, ICE is ideal for fast prioritization.

Formula

ICE Score = Impact × Confidence × Ease

Factors (All 1-10 Scale)

Factor Question

Impact How much will this move the metric?

Confidence How sure are we this will work?

Ease How easy is this to implement?

Example

Feature: One-click checkout

Impact: 9 (directly increases conversion) Confidence: 7 (similar features work elsewhere) Ease: 4 (requires payment integration work)

ICE = 9 × 7 × 4 = 252

ICE vs RICE

Aspect RICE ICE

Complexity More detailed Simpler

Reach consideration Explicit Implicit in Impact

Effort Person-months 1-10 Ease scale

Best for Data-driven teams Fast decisions

WSJF (Weighted Shortest Job First)

SAFe framework optimizing for economic value delivery.

Formula

WSJF = Cost of Delay / Job Size

Cost of Delay Components

Cost of Delay = User Value + Time Criticality + Risk Reduction

Component Question Scale

User Value How much do users/business want this? 1-21 (Fibonacci)

Time Criticality Does value decay over time? 1-21

Risk Reduction Does this reduce risk or enable opportunities? 1-21

Job Size Relative effort compared to other items 1-21

Time Criticality Guidelines

Score Situation

21 Must ship this quarter or lose the opportunity

13 Competitor pressure, 6-month window

8 Customer requested, flexible timeline

3 Nice to have, no deadline

1 Can wait indefinitely

Example

Feature: GDPR compliance update

User Value: 8 (required for EU customers) Time Criticality: 21 (regulatory deadline) Risk Reduction: 13 (avoids fines) Job Size: 8 (medium complexity)

Cost of Delay = 8 + 21 + 13 = 42 WSJF = 42 / 8 = 5.25

MoSCoW Method

Qualitative prioritization for scope management.

Categories

Priority Meaning Guideline

Must Have Non-negotiable for release ~60% of effort

Should Have Important but not critical ~20% of effort

Could Have Nice to have if time permits ~20% of effort

Won't Have Explicitly out of scope Documented

Application Rules

  • Must Have items alone should deliver a viable product

  • Should Have items make product competitive

  • Could Have items delight users

  • Won't Have prevents scope creep

Template

Release 1.0 MoSCoW

Must Have (M)

  • User authentication
  • Core data model
  • Basic CRUD operations

Should Have (S)

  • Search functionality
  • Export to CSV
  • Email notifications

Could Have (C)

  • Dark mode
  • Keyboard shortcuts
  • Custom themes

Won't Have (W)

  • Mobile app (Release 2.0)
  • AI recommendations (Release 2.0)
  • Multi-language support (Release 3.0)

Kano Model

Categorize features by customer satisfaction impact.

Categories

Type Absent Present Example

Must-Be Dissatisfied Neutral Login works

Performance Dissatisfied Satisfied Fast load times

Delighters Neutral Delighted AI suggestions

Indifferent Neutral Neutral About page design

Reverse Satisfied Dissatisfied Forced tutorials

Kano Survey Questions

For each feature, ask two questions:

  • "How would you feel if this feature was present?"

  • "How would you feel if this feature was absent?"

Answer options: Like it, Expect it, Neutral, Can tolerate, Dislike

Framework Selection Guide

Situation Recommended Framework

Data-driven team with metrics RICE

Fast startup decisions ICE

SAFe/Agile enterprise WSJF

Fixed scope negotiation MoSCoW

Customer satisfaction focus Kano

Strategic alignment Value vs. Effort Matrix

Common Pitfalls

Pitfall Mitigation

Gaming the scores Calibrate as a team regularly

Ignoring qualitative factors Use frameworks as input, not gospel

Analysis paralysis Set time limits on scoring sessions

Inconsistent scales Document and share scoring guidelines

Practical Tips

  • Calibrate together: Score several items as a team to align understanding

  • Revisit regularly: Priorities shift—rescore quarterly

  • Document assumptions: Why did you give that Impact score?

  • Combine frameworks: Use ICE for quick triage, RICE for final decisions

Related Skills

  • product-strategy-frameworks

  • Strategic context for prioritization

  • okr-kpi-patterns

  • Connect priorities to measurable goals

  • requirements-engineering

  • Detailed specs for prioritized items

References

  • RICE Deep Dive

  • WSJF Calculator

Version: 1.0.0 (January )

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

General

ui-components

No summary provided by upstream source.

Repository SourceNeeds Review
General

responsive-patterns

No summary provided by upstream source.

Repository SourceNeeds Review
General

domain-driven-design

No summary provided by upstream source.

Repository SourceNeeds Review