meta-prompter

Prompt evaluation and optimization using the meta-prompter-mcp CLI. Returns <OPTIMIZED_PROMPT> for the caller to execute.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "meta-prompter" with this command: npx skills add delexw/claude-code-misc/delexw-claude-code-misc-meta-prompter

Meta-Prompter

Prompt evaluation and optimization using the meta-prompter-mcp CLI. Returns <OPTIMIZED_PROMPT> for the caller to execute.

Inputs

Raw arguments: $ARGUMENTS

Infer from the arguments:

  • CONTEXT: the task context or prompt to optimize

  • OUT_DIR: output directory, or .implement-assets/meta-prompter if not provided

Session

  • If starting fresh: create SESSION_ID = sess-YYYYMMDD-HHMMSS and use it for all files.

Execution

See references/rules.md for CLI usage, environment variables (model configuration), and error handling.

Execute all steps A through E:

A) Evaluate

  • If <prompt_eval> tag doesn't exist:

  • Run: npx meta-prompter-mcp "CONTEXT" via Bash, with --model flag per references/rules.md model resolution

  • Save the JSON output to <prompt_eval> and include:

  • "original_prompt": "CONTEXT"

  • Otherwise, skip to Clarify.

B) Clarify

  • If <clarification_answers> tag is missing or incomplete:

  • Read the 3 questions from <prompt_eval>

  • Attempt to answer all questions based on context

  • If you don't have enough context to answer:

  • Use AskUserQuestion tool to ask each question

  • Save the questions and answers to <clarification_answers> tag as: { "answers": [ {"q": "<question1>", "answer": "<user answer>"}, {"q": "<question2>", "answer": "<user answer>"}, {"q": "<question3>", "answer": "<user answer>"} ] }

C) Build PROMPT

  • Proceed only when all 3 answers are present and clear.

  • Build PROMPT:

  • Base on rewrite from evaluation result

  • Append a short Clarifications section from the 3 answers (concise bullet points).

D) Re-evaluate and gate

  • Think the re-evaluation process carefully

  • Re-run npx meta-prompter-mcp "<built PROMPT>" via Bash (with --model flag per references/rules.md model resolution)

  • Overwrite <prompt_eval> with this latest evaluation JSON result (preserve "original_prompt" and, if present, "contextual_prompt" ).

  • When global < 8 :

  • STOP execution immediately.

  • Do NOT invent questions. Use the 3 questions returned by this re-evaluation.

  • Redo B) Clarify, redo C) Build PROMPT, and re-run this step (D).

  • Repeat until global >= 8 , then proceed using the <OPTIMIZED_PROMPT>.

  • Max 3 iterations. If global < 8 after 3 attempts, use the best-scoring prompt as <OPTIMIZED_PROMPT> and note the score.

E) Return

  • If OUT_DIR is provided:

  • Run mkdir -p OUT_DIR via Bash

  • Use the Write tool to save <OPTIMIZED_PROMPT> to OUT_DIR/soul.md

  • Return <OPTIMIZED_PROMPT> to the caller for execution

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Coding

figma-reader

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

forge

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

pir

No summary provided by upstream source.

Repository SourceNeeds Review