Shinka Inspect Skill
Extract the strongest programs from a Shinka run and package them into a context file that coding agents can load directly.
When to Use
Use this skill when:
- A run already produced a results directory and SQLite database
- You want to inspect top-performing programs before launching the next batch
- You want a compact context artifact instead of manually browsing the DB
Do not use this skill when:
- You still need to scaffold a task (
shinka-setup) - You need to run evolution batches (
shinka-run)
What it does
- Uses
shinka.utils.load_programs_to_dfto read program records - Ranks programs by
combined_score - Selects top-
kcorrect programs (fallback to top-koverall if no correct rows) - Writes one Markdown bundle with metadata, ranking table, feedback, and code snippets
Workflow
- Confirm run artifacts exist
ls -la <results_dir>
- Generate context bundle
python skills/shinka-inspect/scripts/inspect_best_programs.py \
--results-dir <results_dir> \
--k 5
- Optional tuning knobs
python skills/shinka-inspect/scripts/inspect_best_programs.py \
--results-dir <results_dir> \
--k 8 \
--max-code-chars 5000 \
--min-generation 10 \
--out <results_dir>/inspect/top_programs.md
- Load output into agent context
- Default output path:
<results_dir>/shinka_inspect_context.md - Use it as the context artifact for next-step mutation planning
CLI Arguments
--results-dir: Path to run directory (or direct DB file path)--k: Number of programs to include (default5)--out: Output markdown path (default under results dir)--max-code-chars: Per-program code truncation cap (default4000)--min-generation: Optional lower bound on generation--include-feedback/--no-include-feedback: Includetext_feedbackblocks
Notes
- Ranking metric is
combined_score. - If no correct rows exist, script falls back to top-score rows and labels fallback in output.
- Script is read-only for run artifacts (writes only the markdown bundle).