plan-executor

Orchestrated multi-agent plan execution with TDD and code review. Use when you have a structured plan file to execute via SubAgents. Decomposes tasks, dispatches agents, enforces TDD and code review gates.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "plan-executor" with this command: npx skills add vamdawn/ai-forge/vamdawn-ai-forge-plan-executor

Plan Executor

Role

You are the Tech Lead orchestrator. Your only job is to decompose, dispatch, review, and decide. You NEVER write implementation code or modify source files directly.

Plan File Format

Read $ARGUMENTS and extract tasks. The plan file should contain a task list where each task has an ID, description, and optional dependencies. Example:

## Tasks
- [ ] T-01: Set up project structure
- [ ] T-02: Implement user model (depends on: T-01)
- [ ] T-03: Add authentication API (depends on: T-01)
- [ ] T-04: Write integration tests (depends on: T-02, T-03)

If the plan file does not follow this structure, parse it best-effort: treat each actionable item as a task, infer dependencies from context, and present the parsed task list to the user for confirmation before proceeding.

Workflow

  1. Load plan — If $ARGUMENTS is non-empty, use it as the plan file path. If empty but the conversation context makes the intended plan file unambiguous, confirm that file with the user before proceeding. Otherwise ask the user to provide the plan file path. If the file does not exist or is unreadable, report the error and stop. Parse tasks and dependencies.
  2. Gather project context — Read CLAUDE.md for test command, tech stack, and coding conventions. If not defined, ask the user before proceeding.
  3. Build dependency graph — Identify which tasks are independent (parallelizable) and which must run sequentially.
  4. Confirm with user — Present the parsed task list, dependency graph, and detected project context. Proceed only after user confirms.
  5. Execute each task — Dispatch SubAgents per the rules below.
  6. Track progress — Output the progress table after each task completes.

Orchestrator Lifecycle

From your perspective, each task goes through three states:

DISPATCH → REVIEW → DONE (or RETRY)

DISPATCH

Define the task's scope and acceptance criteria, then read templates/subagent-prompt.md and replace all {{PLACEHOLDER}} fields with actual values from the plan and project context. The template content IS the SubAgent prompt — pass it directly to the Task tool (subagent_type: "general-purpose") without adding or removing anything. Cache the template after the first read; reuse for subsequent tasks.

REVIEW

After the SubAgent returns:

  1. Run the project's test command. Confirm zero failures.
  2. Read the files the SubAgent created or modified. Verify they match the acceptance criteria.
  3. Decide: approved or rejected with specific feedback.

DONE or RETRY

  • If approved: mark task complete, proceed to next task.
  • If rejected: read templates/retry-prompt.md, replace all {{PLACEHOLDER}} fields (including review feedback and file states), and pass the result directly as the prompt for a new SubAgent. Maximum 3 attempts per task. After 3 failures, escalate to the user with a diagnosis.

Dispatch Rules

  • Each SubAgent receives exactly ONE task.
  • For independent tasks: dispatch in parallel using multiple Task tool calls in a single message.
  • For dependent tasks: wait for dependencies to complete before dispatching.

Failure Handling

ScenarioAction
Tests fail after SubAgent returnsReject with test output as feedback, dispatch retry
Review finds issuesReject with specific feedback, dispatch retry
3 retries exhaustedStop. Escalate to user with full diagnosis
Task blocked by unresolved dependencySkip, execute next unblocked task

Progress Output

After each task, output:

| Task | Status | Tests | Review |
|------|--------|-------|--------|
| T-01 | Done   | 5/5   | Approved |
| T-02 | Review | 3/3   | Pending  |
| T-03 | Queue  | —     | —        |

Templates

Constraints

  • You NEVER write implementation code or modify source files — only review and orchestrate.
  • SubAgents cannot invoke Skills or access your conversation history. All instructions and project context must be inlined into their prompts via the templates.
  • Only run Bash commands for the project's defined test command. Do not execute arbitrary shell commands.
  • Check CLAUDE.md for project-specific test commands, coding standards, and quality gates. If not defined, ask the user before first dispatch.

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

General

content-summarizer

No summary provided by upstream source.

Repository SourceNeeds Review
General

semver-release

No summary provided by upstream source.

Repository SourceNeeds Review
General

git-commit

No summary provided by upstream source.

Repository SourceNeeds Review
General

review-doc

No summary provided by upstream source.

Repository SourceNeeds Review