gpt-5-4-prompting

Use this skill when codex:codex-rescue needs to ask Codex or another GPT-5.4-based workflow for help.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "gpt-5-4-prompting" with this command: npx skills add openai/codex-plugin-cc/openai-codex-plugin-cc-gpt-5-4-prompting

GPT-5.4 Prompting

Use this skill when codex:codex-rescue needs to ask Codex or another GPT-5.4-based workflow for help.

Prompt Codex like an operator, not a collaborator. Keep prompts compact and block-structured with XML tags. State the task, the output contract, the follow-through defaults, and the small set of extra constraints that matter.

Core rules:

  • Prefer one clear task per Codex run. Split unrelated asks into separate runs.

  • Tell Codex what done looks like. Do not assume it will infer the desired end state.

  • Add explicit grounding and verification rules for any task where unsupported guesses would hurt quality.

  • Prefer better prompt contracts over raising reasoning or adding long natural-language explanations.

  • Use XML tags consistently so the prompt has stable internal structure.

Default prompt recipe:

  • <task> : the concrete job and the relevant repository or failure context.

  • <structured_output_contract> or <compact_output_contract> : exact shape, ordering, and brevity requirements.

  • <default_follow_through_policy> : what Codex should do by default instead of asking routine questions.

  • <verification_loop> or <completeness_contract> : required for debugging, implementation, or risky fixes.

  • <grounding_rules> or <citation_rules> : required for review, research, or anything that could drift into unsupported claims.

When to add blocks:

  • Coding or debugging: add completeness_contract , verification_loop , and missing_context_gating .

  • Review or adversarial review: add grounding_rules , structured_output_contract , and dig_deeper_nudge .

  • Research or recommendation tasks: add research_mode and citation_rules .

  • Write-capable tasks: add action_safety so Codex stays narrow and avoids unrelated refactors.

How to choose prompt shape:

  • Use built-in review or adversarial-review commands when the job is reviewing local git changes. Those prompts already carry the review contract.

  • Use task when the task is diagnosis, planning, research, or implementation and you need to control the prompt more directly.

  • Use task --resume-last for follow-up instructions on the same Codex thread. Send only the delta instruction instead of restating the whole prompt unless the direction changed materially.

Working rules:

  • Prefer explicit prompt contracts over vague nudges.

  • Use stable XML tag names that match the block names from the reference file.

  • Do not raise reasoning or complexity first. Tighten the prompt and verification rules before escalating.

  • Ask Codex for brief, outcome-based progress updates only when the task is long-running or tool-heavy.

  • Keep claims anchored to observed evidence. If something is a hypothesis, say so.

Prompt assembly checklist:

  • Define the exact task and scope in <task> .

  • Choose the smallest output contract that still makes the answer easy to use.

  • Decide whether Codex should keep going by default or stop for missing high-risk details.

  • Add verification, grounding, and safety tags only where the task needs them.

  • Remove redundant instructions before sending the prompt.

Reusable blocks live in references/prompt-blocks.md. Concrete end-to-end templates live in references/codex-prompt-recipes.md. Common failure modes to avoid live in references/codex-prompt-antipatterns.md.

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Coding

codex-cli-runtime

No summary provided by upstream source.

Repository SourceNeeds Review
258-openai
Coding

Hybrid Deep Search

Hybrid Deep Search - Intelligent routing between Brave API (free, fast) and OpenAI Codex (deep analysis, paid). Auto-selects optimal search strategy based on...

Registry SourceRecently Updated
6640Profile unavailable
Coding

Gitai - Git Commit Automation

Boost developer productivity with Gitai: An AI-powered Git automation tool that analyzes code changes and generates semantic Conventional Commits instantly. Supports Node.js, Python, Java, Go, PHP, and more. Compatible with OpenAI, Anthropic, and Groq.

Registry SourceRecently Updated
3.2K1Profile unavailable
Coding

develop-web-game

No summary provided by upstream source.

Repository SourceNeeds Review
1.1K-openai