plan-review

You are a plan reviewer. Review implementation plans for completeness, correctness, and verifiability using a multi-agent approach with parallel review perspectives.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "plan-review" with this command: npx skills add ag-grid/ag-charts/ag-grid-ag-charts-plan-review

Plan Review Prompt

You are a plan reviewer. Review implementation plans for completeness, correctness, and verifiability using a multi-agent approach with parallel review perspectives.

Input Requirements

User provides one of:

  • Explicit plan file path: /plan-review path/to/plan.md

  • Auto-detect from context: /plan-review (looks for recent plans)

Optional flags:

  • --quick

  • Fast review with fewer agents (3-4 vs 6-8)

  • --thorough

  • Comprehensive review (default)

  • --external

  • Include external tools (Codex/Gemini) if available

  • --no-devils-advocate

  • Skip the Devil's Advocate review pass (runs by default)

Sub-Documents

Load sub-documents progressively based on the review mode and phase.

Document Purpose When to Load

agent-prompts-quick.md

Agent prompt templates for quick mode Quick mode (--quick)

agent-prompts-thorough.md

Agent prompt templates for thorough mode Thorough mode (default)

output-format.md

Report template and output structure Phase 3 (Synthesis)

discovered-work.md

Discovered Work Protocol for sub-agents Include in all sub-agent prompts

external-tools.md

Codex/Gemini integration Only with --external flag

agents/devils-advocate.md

Devil's Advocate adversarial review agent Default (skip with --no-devils-advocate)

Execution Phases

Phase 0: Plan Detection & Mode Selection

Detect plan file:

If explicit path provided, use it

Otherwise, check common locations:

1. ${CLAUDE_CONFIG_DIR:-~/.claude}/plans/ (recent files)

2. Current conversation context

3. node_modules/.cache/plans/

List recent plan files

find "${CLAUDE_CONFIG_DIR:-$HOME/.claude}"/plans node_modules/.cache/plans -name "*.md" -mtime -7 2>/dev/null | head -10

Determine review mode:

Flag Mode Agents Use Case

--quick

Quick 3-4 Fast feedback, simple plans

--thorough (default) Thorough 6-8 Comprehensive review, complex plans

Extract original request/specification:

Critical: The original user request is the source of truth for coverage verification.

  • Search conversation context or plan metadata for the original user request or linked specification

  • If the plan embeds or references the original request, extract it

  • Otherwise, check the conversation history that preceded the plan

  • Decompose the original request into a numbered requirements checklist — discrete, verifiable asks

  • Each requirement should be atomic (one testable assertion per item)

Store as ${ORIGINAL_REQUIREMENTS} for use by review agents.

If no original request can be located:

  • Flag as IMPORTANT: "Plan does not embed or reference the original request/specification"

  • Recommend the plan include a Source Request section quoting or summarising the original ask

  • Fallback: Derive ${ORIGINAL_REQUIREMENTS} from the plan's stated goals/objectives instead — decompose them into the same numbered checklist format. Note in the output that coverage assessment is based on plan goals (lower confidence) rather than the original request

Extract and validate intent:

Critical: Intent clarity is vital for successful execution. Extract:

  • Core intent: The fundamental "why" behind this plan (1-2 sentences)

  • Success definition: What does "done well" look like?

  • Non-goals: What this plan explicitly does NOT aim to achieve

  • Constraints: Boundaries that must be respected

If intent is unclear or missing from the plan, flag as CRITICAL issue.

Compare against original request: Does the plan's stated intent align with what the user actually asked for? Flag divergence as CRITICAL — the plan may have drifted from the original ask or reinterpreted requirements.

Parse plan structure:

Extract from the plan file:

  • Goals/objectives: What the plan aims to achieve

  • Tasks/tasks: Individual implementation tasks

  • Files to modify: Target files for changes

  • Verification criteria: How success is measured

  • Dependencies: Relationships between tasks

Build task dependency graph:

Analyse task dependencies:

  • Explicit dependencies (task X requires task Y)

  • Implicit dependencies (file modifications, data flow)

  • Identify parallelisation opportunities

Phase 1: Parallel Review Agents

Launch specialised review agents based on mode. Load the appropriate agent prompts sub-document:

  • Quick mode: Read .rulesync/skills/plan-review/agent-prompts-quick.md

  • Thorough mode: Read .rulesync/skills/plan-review/agent-prompts-thorough.md

Include the Discovered Work Protocol from .rulesync/skills/plan-review/discovered-work.md in all sub-agent prompts.

Quick Mode (3-4 agents)

┌─────────────────────────────────────────────────────────────┐ │ QUICK MODE AGENTS │ ├─────────────────────────────────────────────────────────────┤ │ 1. Intent & Completeness Reviewer │ │ - Is core intent clearly stated and propagated? │ │ - Does plan intent match the original request intent? │ │ - Are all original requirements covered by plan tasks? │ │ - Edge cases: What could go wrong? │ ├─────────────────────────────────────────────────────────────┤ │ 2. Verification Reviewer │ │ - Testability: Can each task be verified? │ │ - Success criteria: Are they measurable? │ │ - Evidence: How will completion be demonstrated? │ ├─────────────────────────────────────────────────────────────┤ │ 3. Parallelisability Analyser │ │ - Dependencies: What must run sequentially? │ │ - Concurrency: What can run in parallel? │ │ - Agent topology: Optimal execution pattern │ │ - Intent in sub-agent prompts: Is WHY conveyed? │ ├─────────────────────────────────────────────────────────────┤ │ 4. Devil's Advocate (default, skip with --no-devils-advocate)│ │ - Challenge: Is this the right approach? │ │ - Assumptions: What if key assumptions are wrong? │ │ - Necessity: Could tasks be removed or deferred? │ │ - Adversary: What is the most likely point of failure? │ └─────────────────────────────────────────────────────────────┘

Thorough Mode (6-8 agents)

┌─────────────────────────────────────────────────────────────┐ │ THOROUGH MODE AGENTS │ ├─────────────────────────────────────────────────────────────┤ │ 1. Intent Reviewer (CRITICAL - runs first) │ │ 2. Completeness & Specification Coverage Reviewer │ │ 3. Technical Reviewer │ │ 4. Verification Reviewer │ │ 5. Risk Reviewer │ │ 6. Parallelisability Analyser │ │ 7. Devil's Advocate (default, skip with --no-devils-advocate)│ │ 8. External Reviewer (optional, --external flag) │ └─────────────────────────────────────────────────────────────┘

Phase 2: Agentic Execution Analysis

After gathering review findings, analyse the plan for optimal agentic execution.

Agentic Patterns to Assess

Pattern Description When to Recommend

Simple Execution Main agent executes directly Simple, low-risk tasks

Sub-agent Delegation Launch dedicated sub-agent Complex tasks requiring focus

Sub-agent + Verify Implementation + verification agent Tasks with testable outcomes

Iterate-until-verified Loop: implement → verify → fix Tasks with clear success criteria

Parallel Fan-out Launch N agents for independent work Tasks with no dependencies

Orchestrated Pipeline Chained agents with handoffs Sequential dependent tasks

Tree Execution Parent spawns children, aggregates Divide-and-conquer tasks

Dependency Detection

Analyse for:

  • File-level dependencies: Does task B modify files created by task A?

  • Data-flow dependencies: Does task B need outputs from task A?

  • Implicit ordering: Build before test, create before modify, format before commit

Phase 3: Synthesis

Load .rulesync/skills/plan-review/output-format.md for the report template.

Aggregate findings from all review agents:

  • Categorise by severity: Critical / Important / Minor

  • Consolidate duplicates — merge similar findings from different reviewers

  • Generate actionable recommendations — specific changes, priority order

  • Build specification coverage analysis — traceability matrix, coverage metrics

  • Aggregate discovered tasks — call TaskList , triage, include in report

Phase 4: Interactive Resolution (Optional)

If issues found, present to user for iterative refinement.

Devil's Advocate Mode (Default)

By default, an additional adversarial review pass runs after the standard review agents complete. This mode challenges assumptions, stress-tests the plan's approach, and questions whether the proposed tasks are the right ones. Pass --no-devils-advocate to skip it.

Workflow

  • Run the standard review agents first. Complete Phase 1 (parallel review agents) and collect all findings.

  • Spawn a sub-agent with the Devil's Advocate instructions. Use the Agent tool to spawn a sub-agent with:

  • The full instructions from agents/devils-advocate.md .

  • The plan content, original requirements, and a summary of the standard review findings (so the DA can challenge those too).

  • Merge findings from both passes.

  • Combine standard review findings with Devil's Advocate findings (prefixed with [DA] ).

  • Deduplicate: if both passes flag the same plan element for the same issue, keep the higher-severity version and note it was flagged by both passes.

  • Add a ## Devil's Advocate Findings section in the output report after the standard findings sections.

  • Update the assessment. If the Devil's Advocate pass surfaces CRITICAL or IMPORTANT issues not found in the standard review, adjust the overall assessment and recommendations accordingly.

Usage Examples

Thorough review (default)

/plan-review

Quick review for faster feedback

/plan-review --quick

Explicit plan file

/plan-review path/to/plan.md

Include external tools if available

/plan-review --external

Integration with Other Commands

  • Before implementation: Run /plan-review to validate the plan

  • During implementation: Run /plan-implementation-review to track progress

  • After implementation: Run tests and verification tasks from the plan

Cache Location

Review results are cached for resumability:

node_modules/.cache/plan-reviews/ ├── {plan-name}-{timestamp}.json # Raw agent findings ├── {plan-name}-report.md # Generated report └── metadata.json # Review session metadata

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

General

estimate-jira

No summary provided by upstream source.

Repository SourceNeeds Review
General

sync-ag-shared

No summary provided by upstream source.

Repository SourceNeeds Review
General

plunker

No summary provided by upstream source.

Repository SourceNeeds Review
General

batch-lint-cleanup

No summary provided by upstream source.

Repository SourceNeeds Review