product-management

Product Management (Jan 2026)

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "product-management" with this command: npx skills add vasilyu1983/ai-agents-public/vasilyu1983-ai-agents-public-product-management

Product Management (Jan 2026)

This skill turns the assistant into an operator, not a lecturer.

Everything here is:

  • Executable: templates, checklists, decision flows

  • Decision-first: measurable outcomes, explicit trade-offs, clear ownership

  • Organized: resources for depth; templates for immediate copy-paste

Modern Best Practices (Jan 2026):

  • Evidence quality beats confidence: label signals strong/medium/weak; write what would change your mind.

  • Outcomes > output: roadmaps are bets with measurable impact and guardrails, not feature inventories.

  • Metrics must be defined (formula + timeframe + data source) to be actionable.

  • Privacy, security, and accessibility are requirements, not afterthoughts.

  • Hybrid decision loops: AI surfaces anomalies, patterns, and forecasts; humans apply context, ethics, and long-term strategy.

  • Accountability: product is often held responsible for business outcomes; confirm the operating model in your org and validate benchmarks with current sources.

  • Portfolio diversification: a common heuristic is 70% core, 20% adjacent, 10% transformational; adapt to strategy and constraints.

When to Use This Skill

Use this skill when the user asks to do real product work, such as:

  • “Create / refine a PRD / spec / business case / 1-pager”

  • “Turn this idea into a roadmap” / “Outcome roadmap for X”

  • “Design a discovery plan / interview script / experiment plan”

  • “Define success metrics / OKRs / metric tree”

  • “Position this product against competitors”

  • “Run a difficult conversation / feedback / 1:1 / negotiation”

  • “Plan a product strategy / vision / opportunity assessment”

Do not use this skill for:

  • Book summaries, philosophy, or general education

  • Long case studies or storytelling

Quick Reference

Task Template Domain Output

Discovery interview customer-interview-template.md

Discovery Interview script with Mom Test patterns

Opportunity mapping opportunity-solution-tree.md

Discovery OST with outcomes, problems, solutions

PMF survey pmf-survey-template.md

Discovery Sean Ellis + NPS + usage survey

Outcome roadmap outcome-roadmap.md

Roadmap Now/Next/Later with outcomes and themes

OKR definition okr-template.md

Metrics 1-3 objectives with 2-4 key results each

Product positioning positioning-template.md

Strategy Competitive alternatives -> value -> segment

Product vision product-vision-template.md

Strategy From→To narrative with 3-5 year horizon

Quarterly review quarterly-product-review.md

Strategy Keep / cut / double-down product audit

Prioritization prioritization-scorecard.md

Prioritization RICE/ICE scoring with kill criteria

Kill criteria kill-criteria-template.md

Prioritization Pre-defined stop conditions per initiative

1:1 meeting 1-1-template.md

Leadership Check-in, progress, blockers, growth

Post-incident debrief a3-debrief.md

Leadership Intent vs actual, root cause, action items

Decision Tree: Choosing the Right Workflow

User needs: [Product Work Type] ├─ Discovery / Validation? │ ├─ Customer insights? → Customer interview template │ ├─ Hypothesis testing? → Assumption test template │ └─ Opportunity mapping? → Opportunity Solution Tree │ ├─ Strategy / Vision? │ ├─ Long-term direction? → Product vision template │ ├─ Market positioning? → Positioning template (Dunford) │ ├─ Big opportunity? → Opportunity assessment │ └─ Amazon-style spec? → PR/FAQ template │ ├─ Planning / Roadmap? │ ├─ Outcome-driven? → Outcome roadmap (Now/Next/Later) │ ├─ Theme-based? → Theme roadmap │ └─ Metrics / OKRs? → Metric tree + OKR template │ ├─ Prioritization / Focus? │ ├─ What to build next? → Prioritization scorecard (RICE/ICE) │ ├─ What to stop? → Kill criteria template + quarterly review │ ├─ Scope too large? → Scope negotiation patterns │ └─ PMF check? → PMF survey + retention curve analysis │ └─ Leadership / Team Ops? ├─ 1:1 meeting? → 1-1 template ├─ Giving feedback? → Feedback template (SBI model) ├─ Post-incident? → A3 debrief ├─ Stakeholder pushback? → Stakeholder management patterns └─ Negotiation? → Negotiation one-sheet (Voss)

Do / Avoid (Jan 2026)

Do

  • Start from the decision: what are we deciding, by when, and with what evidence.

  • Define metrics precisely (formula + timeframe + data source) and add guardrails.

  • Use discovery to de-risk value before building; prioritize by evidence, not opinions.

  • Write “match vs ignore” competitive decisions, not feature grids.

Avoid

  • Roadmap theater (shipping lists) without outcomes and learning loops.

  • Vanity KPIs (raw signups, impressions) without activation/retention definitions.

  • "Build-first validation" (shipping MVPs without falsifiable hypotheses).

  • Collecting customer data without purpose limitation, retention, and access controls.

  • Building for engineering elegance instead of user value (technical founder trap).

  • Feature creep without kill criteria (every feature should have a pre-defined stop condition).

  • Saying "yes" to stakeholder requests without trade-off analysis.

  • Measuring PMF once instead of continuously across segments.

Prioritization & Saying No

The most common founder-PM failure: building everything, killing nothing, and running out of time before impact.

Prioritization Frameworks

Framework Formula / Method Best For Watch For

RICE (Reach x Impact x Confidence) / Effort Comparing features with data Gaming confidence scores

ICE Impact x Confidence x Ease Quick gut-check prioritization Over-simplification

Opportunity Scoring Importance x (Importance - Satisfaction) Discovery-driven, JTBD-aligned Requires user research data

Cost of Delay Value per unit time / Duration Time-sensitive decisions Harder to estimate accurately

Weighted Shortest Job First (WSJF) Cost of Delay / Job Size SAFe/Lean, flow optimization Requires calibrated estimates

Pick one. Use it consistently. The framework matters less than the discipline of scoring everything the same way.

Kill Criteria

Every initiative should have pre-defined conditions for stopping:

  • Usage threshold: If <X% of target users adopt within Y weeks, stop.

  • Cost ceiling: If development exceeds X hours/dollars, pause and re-evaluate.

  • Time limit: If not shipped within X weeks, kill or radically descope.

  • Metric guardrail: If [guardrail metric] degrades by >X%, roll back.

Use assets/prioritization/kill-criteria-template.md to define these before starting.

Feature Bridge Migration

When replacing an existing feature with a new one, don't hard-kill the old feature. Use a bridge migration pattern to prevent user loss.

Bridge mode: Run both old and new features simultaneously. Route users to the new experience by default but keep the old path accessible (via link, fallback, or settings toggle).

Substitution-based kill rule:

  • Define the absorption metric: % of old-feature users who now use the new feature for the same job.

  • Set the kill threshold: new feature absorbs ≥80% of old-feature users.

  • Set the duration: threshold must hold for 14 consecutive days with no retention regression.

  • Only kill the old feature when all three conditions are met.

BRIDGE MIGRATION SEQUENCE:

  1. Ship new feature alongside old feature
  2. Default new users to new experience
  3. Migrate existing users gradually (progressive rollout)
  4. Monitor: absorption rate, retention by cohort, support tickets
  5. Old feature absorbs ≥80% for 14 days + no retention drop? ├─ Yes → Kill old feature, remove code └─ No → Investigate gaps, iterate new feature, extend bridge

When NOT to bridge: Security vulnerabilities, compliance requirements, or features with near-zero usage (<1% MAU). These can be killed directly with notice.

Scope Negotiation

When stakeholders push for more scope:

  • Reframe as trade-offs: "We can add X if we cut Y — which matters more?"

  • Anchor on outcomes: "The goal is [metric]. Does this addition move it?"

  • Offer phased delivery: "V1 without this; measure; add in V2 if data supports it."

  • Document non-goals explicitly in every spec.

"What to Stop Doing" Quarterly Review

Every quarter, review the product with assets/strategy/quarterly-product-review.md :

  • Which features have <5% usage? → Candidate for removal

  • Which initiatives produced no measurable outcome? → Stop or pivot

  • Which ongoing costs (maintenance, support) exceed their value? → Sunset

  • What are you doing "because we always have" but nobody asked for? → Question

For detailed prioritization patterns and worked examples: see references/prioritization-frameworks.md .

Product-Market Fit Measurement

PMF is not a binary event. It's a signal you measure across multiple dimensions.

Sean Ellis Test

Survey users: "How would you feel if you could no longer use [product]?"

  • Very disappointed: Target >40% for PMF signal

  • Somewhat disappointed: Useful but not dependent

  • Not disappointed: Not finding value

Use assets/discovery/pmf-survey-template.md for the full survey (combines Sean Ellis + NPS + usage questions).

Retention Curve Analysis

  • Plot cohort retention over time (weekly or monthly depending on product cadence)

  • Flattening curve = PMF signal (users who stay, stay)

  • Declining curve = No PMF (even retained users eventually leave)

  • Segment by ICP: you may have PMF in one segment but not another

Engagement Scoring

Define activation precisely (formula + timeframe + data source):

  • What actions constitute "activated"? (not just signed up)

  • What's the activation window? (first 7 days, first 14 days?)

  • What engagement depth separates power users from casual?

Feature Audit

Periodically audit feature usage to identify what to keep, improve, or remove:

  • Top 20% features by usage → invest, polish

  • Middle 60% → maintain, don't expand

  • Bottom 20% → candidate for removal or redesign

  • Features with high support cost relative to usage → redesign or sunset

Segmented PMF

PMF varies by segment. Measure separately for:

  • ICP vs non-ICP customers

  • Free vs paid users

  • Self-serve vs sales-assisted

  • By company size, industry, or geography

For detailed PMF measurement methodology: see references/pmf-measurement.md .

Stakeholder Management

Founders manage board members, investors, early customers, co-founders, and (eventually) team leads — often without formal PM training.

Key patterns:

  • Board / investors: Update monthly with metrics + decisions + asks. Use narrative format, not slide decks. Lead with "what we learned" not "what we shipped."

  • Early customers: They are partners, not just users. Share roadmap intent (not commitments). Ask for input on priorities, not feature requests.

  • Co-founder alignment: Weekly sync on priorities. Disagree and commit. Document decisions.

  • Saying no to stakeholders: "We're not doing X because [reason tied to strategy]. Here's what we're doing instead and why."

For detailed stakeholder management patterns: see references/stakeholder-management.md .

What Good Looks Like

  • Evidence: 5–10 real user touchpoints or equivalent primary data for material bets.

  • Scope: clear non-goals and acceptance criteria that can be tested.

  • Learning: post-launch review with metric deltas, guardrail impact, and next decision.

PRDs and Specs

For PRDs/specs and writing-quality requirements, use the templates in ../docs-ai-prd/ :

  • PRD templates: ../docs-ai-prd/assets/prd/prd-template.md and ../docs-ai-prd/assets/prd/ai-prd-template.md

Optional: AI / Automation

Use only when explicitly requested and policy-compliant.

  • AI system lifecycle: assets/ai/ai-lifecycle-template.md

  • Agentic workflow docs: assets/ai/agentic-ai-orchestration.md

  • AI product patterns: references/ai-product-patterns.md

Navigation

Resources

  • references/discovery-best-practices.md

  • references/roadmap-patterns.md

  • references/delivery-best-practices.md

  • references/strategy-patterns.md

  • references/positioning-patterns.md

  • references/data-product-best-practices.md

  • references/interviewing-patterns.md

  • references/metrics-best-practices.md

  • references/leadership-decision-frameworks.md

  • references/operational-guide.md

  • references/prioritization-frameworks.md

  • references/pmf-measurement.md

  • references/stakeholder-management.md

  • data/sources.json

Templates

  • Discovery: assets/discovery/customer-interview-template.md, assets/discovery/assumption-test-template.md, assets/discovery/opportunity-solution-tree.md, assets/discovery/pmf-survey-template.md

  • Prioritization: assets/prioritization/prioritization-scorecard.md, assets/prioritization/kill-criteria-template.md

  • Strategy/Vision: assets/strategy/product-vision-template.md, assets/strategy/opportunity-assessment.md, assets/strategy/positioning-template.md, assets/strategy/PRFAQ-template.md, assets/strategy/quarterly-product-review.md

  • Data: assets/data/data-product-canvas.md

  • Roadmaps: assets/roadmap/outcome-roadmap.md, assets/roadmap/theme-roadmap.md

  • Metrics: assets/metrics/metric-tree.md, assets/metrics/okr-template.md

  • Ops/Leadership: assets/ops/1-1-template.md, assets/ops/feedback-template.md, assets/ops/a3-debrief.md, assets/ops/negotiation-one-sheet.md

Related Skills

  • ../docs-ai-prd/SKILL.md — PRD, stories, and prompt/playbook templates

  • ../software-architecture-design/SKILL.md — System design guidance for specs and PRDs

  • ../software-frontend/SKILL.md — UI implementation considerations for product specs

  • ../software-backend/SKILL.md — Backend/API implications of product decisions

  • ../startup-growth-playbooks/SKILL.md — PLG case studies for activation design

Operational Guide

See references/operational-guide.md for detailed patterns, template walkthroughs, example flows, and execution checklists. Keep SKILL.md as the navigation hub; use assets/ when producing artifacts.

External Resources

See data/sources.json for official frameworks (Lean Startup, OST, PR/FAQ, OKRs) and AI/LLM safety references.

Use the quick reference and decision tree above to choose a template, then follow the operational guide for depth.

Trend Awareness Protocol

IMPORTANT: When users ask recommendation questions about product management tools, frameworks, or practices, use a web search tool to check current trends before answering. If web search is unavailable, use data/sources.json and state clearly what you verified vs assumed.

Trigger Conditions

  • "What's the best tool for [roadmapping/product analytics/discovery]?"

  • "What should I use for [OKRs/metrics/customer feedback]?"

  • "What's the latest in product management?"

  • "Current best practices for [discovery/roadmaps/prioritization]?"

  • "Is [framework/tool] still relevant in 2026?"

  • "[Linear] vs [Jira] vs [other]?" or "[Amplitude] vs [Mixpanel]?"

  • "Best approach for [AI product management/agentic products]?"

Required Searches

  • Search: "product management best practices 2026"

  • Search: "[specific tool] vs alternatives 2026"

  • Search: "product management trends January 2026"

  • Search: "[discovery/roadmap/OKR] frameworks 2026"

What to Report

After searching, provide:

  • Current landscape: What PM tools/frameworks are popular NOW

  • Emerging trends: New tools, methods, or patterns gaining traction

  • Deprecated/declining: Frameworks/tools losing relevance

  • Recommendation: Based on fresh data, not just static knowledge

Example Topics (verify with fresh search)

  • Product management tools (Linear, Productboard, Notion, Coda)

  • Analytics platforms (Amplitude, Mixpanel, PostHog)

  • Discovery and research tools (Maze, UserTesting, Dovetail)

  • Roadmapping approaches (outcome-based, theme-based, now/next/later)

  • AI product management patterns

  • Prioritization frameworks (RICE, ICE, opportunity scoring)

  • OKR and metrics tools

Fact-Checking

  • Use web search/web fetch to verify current external facts, versions, pricing, deadlines, regulations, or platform behavior before final answers.

  • Prefer primary sources; report source links and dates for volatile information.

  • If web access is unavailable, state the limitation and mark guidance as unverified.

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Automation

marketing-visual-design

No summary provided by upstream source.

Repository SourceNeeds Review
Automation

startup-idea-validation

No summary provided by upstream source.

Repository SourceNeeds Review
Automation

software-architecture-design

No summary provided by upstream source.

Repository SourceNeeds Review
Automation

marketing-leads-generation

No summary provided by upstream source.

Repository SourceNeeds Review