openai-deep-research-skill

Execute multi-step deep research with the OpenAI Responses API, including question decomposition, evidence gathering with web search, contradiction tracking, and final cited report synthesis. Use when Codex must investigate complex or high-stakes topics (market analysis, policy tracking, technical due diligence, vendor comparison, risk assessment) and deliver structured artifacts (`plan.json`, `findings.json`, `report.md`) rather than ad-hoc answers.

Safety Notice

This listing is from the official public ClawHub registry. Review SKILL.md and referenced scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "openai-deep-research-skill" with this command: npx skills add guanglechen/openai-deep-research-skill

OpenAI Deep Research

Overview

Run a deterministic research workflow that separates planning, evidence collection, and report synthesis. Generate reusable research artifacts under an output directory for auditability and iteration.

Workflow

  1. Define research scope.
  2. Run the script to generate plan, findings, and report artifacts.
  3. Evaluate report quality with the checklist.
  4. Rerun with adjusted depth/model/tool settings when gaps remain.

Quick Start

Install dependencies:

cd openai-deep-research-skill
python3 -m pip install -r scripts/requirements.txt

Run a real research job:

python3 scripts/deep_research.py "中国AI Agent市场2026年商业化路径" \
  --language zh-CN \
  --depth 6 \
  --research-depth deep \
  --max-total-output-tokens 20000 \
  --parallel 3

Run a local dry-run without API calls:

python3 scripts/deep_research.py "sample topic" --dry-run

Runtime Inputs

Set OPENAI_API_KEY before running real jobs. Use OPENAI_BASE_URL only when routing through a compatible gateway.

Tune key flags:

  • --depth: Control breadth of decomposition (2-12).
  • --research-depth: Control per-question evidence depth (shallow|standard|deep).
  • --parallel: Control concurrent evidence runs (1-8).
  • --planner-model: Choose planning model.
  • --research-model: Choose evidence model.
  • --writer-model: Choose synthesis model.
  • --planner-max-output-tokens: Cap planner response size.
  • --research-max-output-tokens: Cap each sub-question research response size.
  • --writer-max-output-tokens: Cap final report synthesis response size.
  • --max-total-output-tokens: Hard limit for estimated run output tokens.
  • --disable-web-search: Disable web tool for internal-data-only runs.
  • --web-tool-type: Override tool type when endpoint uses a non-default web-search tool name.

Artifact Contract

Write one run directory per execution: outputs/<timestamp>-<topic-slug>/. Produce these files:

  • run_meta.json: runtime parameters and metadata.
  • plan.json: normalized sub-question plan.
  • plan_raw.txt: raw planner model output.
  • findings.json: per-question evidence summaries.
  • research_raw.json: raw responses per sub-question.
  • report.md: final cited report.

Quality Gate

Apply all checks before accepting report.md:

  1. Verify each sub-question has explicit evidence or explicit gap notes.
  2. Verify source links are absolute URLs and point to relevant content.
  3. Verify contradictory evidence is surfaced in Contradictions and Uncertainty.
  4. Verify recommendation statements are specific and actionable.
  5. Verify weak-confidence sections are marked clearly.
  6. Verify all required top-level sections exist in Markdown (Executive Summary, Key Findings, Evidence by Sub-question, Contradictions and Uncertainty, Recommendations, Sources).

Use references/research-quality.md for scoring rubric and iteration guidance.

Troubleshooting

If execution fails with missing package errors, install dependencies from scripts/requirements.txt. If JSON parsing fails, rerun with the same topic and lower --depth, then inspect plan_raw.txt or research_raw.json. If web-search tool type is rejected, pass a compatible value via --web-tool-type or disable web search.

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Research

Gougoubi Arena Trade

Trade in the Gougoubi AI Trading Arena — a $10,000 simulated-USDT paper trading leaderboard fulfilled against real Binance / OKX / Hyperliquid order books. A...

Registry SourceRecently Updated
Research

Thinkdeep

Structured reasoning protocol for Claude — forces step-by-step analysis, self-critique, and confidence scoring before answering. Reduces wrong answers and ha...

Registry SourceRecently Updated
Research

股票实时行情分析器

A股/港股实时行情查询、基本面分析、深度报告生成与邮件发送一体化工具。触发场景:(1) 用户询问股票价格、市值、PE/PB等数据;(2) 用户要求分析某只或多只股票;(3) 用户要求生成股票分析报告;(4) 用户要求通过邮件发送股票报告。支持AkShare实时行情、聚宽基本面数据、QQ邮箱/Gmail发送。

Registry SourceRecently Updated
260Profile unavailable
Research

Keep 健康记录

Use when users are stating or logging their own health data to Keep or Keep App rather than asking for advice, analysis, or general chat, including weight, b...

Registry SourceRecently Updated
300Profile unavailable