skill-review

Match user's language: Respond in the same language the user uses.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "skill-review" with this command: npx skills add psylch/better-skills@skill-publish

Skill Review

Language

Match user's language: Respond in the same language the user uses.

Overview

Review a agent skill by combining automated validation with analytical improvement suggestions. Produces a graded report, identifies quality issues, and can interactively apply fixes.

How It Works

  • Identify the skill to review

  • Run automated validation checks with validate.py

  • Extract a structured profile with analyze.sh

  • Read the skill content and compare against improvement patterns

  • Present findings: grade, issues, and prioritized suggestions

  • Interactively apply improvements if the user agrees

Dialogue Flow

Progress:

  • Step 1: Identify the skill

  • Step 2: Automated validation

  • Step 3: Profile extraction

  • Step 4: Deep analysis

  • Step 5: Present findings

  • Step 6: Interactive improvement

Step 1: Identify the Skill

Ask the user for the skill directory path. Auto-detect if the current working directory contains a SKILL.md. Accept absolute or relative paths.

Step 2: Automated Validation

Run the validator to get a graded report:

python3 {SKILL_DIR}/scripts/validate.py run --path <skill-path>

Where {SKILL_DIR} is the directory containing this SKILL.md file.

For detailed output with fix suggestions:

python3 {SKILL_DIR}/scripts/validate.py run --path <skill-path> --format detailed

For strict mode (warnings treated as failures):

python3 {SKILL_DIR}/scripts/validate.py run --path <skill-path> --strict

The script outputs JSON to stdout:

{ "status": "ok", "path": "/path/to/skill", "score": {"total": 22, "pass": 18, "warn": 3, "fail": 1}, "grade": "B", "checks": [ {"id": "skill_md_exists", "category": "structure", "severity": "pass", "message": "SKILL.md found"} ], "hint": "18/22 checks passed, 3 warnings, 1 failure." }

Step 3: Profile Extraction

Run the analyzer to get a structured skill profile:

bash {SKILL_DIR}/scripts/analyze.sh analyze <skill-path>

The script outputs JSON with quantitative facts about the skill: level, section headings, script inventory, feature flags, unfinished-placeholder count, etc.

Step 4: Deep Analysis

Read the target skill's SKILL.md file directly (using the Read tool) to understand its full content, purpose, and workflow. If the skill has scripts or references, read key files as needed.

Then read references/improvement_patterns.md to load the knowledge base of common improvement patterns.

Step 5: Present Findings

Format the report for the user in this order:

  • Grade and summary — Show the letter grade (A/B/C/D/F) and score line from validation

  • Failures — List all fail severity checks with their fix suggestions

  • Warnings — List all warn severity checks

  • Improvement suggestions — Beyond pass/fail checks, compare the skill against improvement patterns and suggest enhancements:

  • What to change and why

  • A concrete before/after example or specific instruction

  • Which file to edit

  • Priority: High (affects functionality/UX), Medium (convention violations), Low (polish)

Present a completion report before asking about fixes:

[Skill Review] Complete!

Target: <skill-name> at <path> Grade: <letter> (<pass>/<total> passed, <warn> warnings, <fail> failures)

Issues found: • <fail count> failure(s) • <warn count> warning(s) • <N> improvement suggestion(s)

If the grade is A or B with no improvement suggestions, congratulate and suggest publishing with skill-publish . If not installed: npx skills add psylch/better-skills@skill-publish -g -y

Step 6: Interactive Improvement

After presenting findings, ask the user which issues and suggestions to address. Options:

  • Fix all — Apply all suggested changes

  • Pick and choose — Let the user select specific items

  • None — Just use the analysis as a reference

For each selected item, make the edit directly (using file editing tools), then confirm the change. After all selected changes, optionally re-run validate.py to show the updated grade.

Check Categories (Automated Validation)

Category What it checks

structure SKILL.md exists, frontmatter present, required fields

naming Kebab-case, length, no consecutive hyphens, matches directory

content Description length, body length, heading structure

paths Referenced files exist, scripts have execute permission

scripts JSON output pattern, preflight subcommand, error handling

security No hardcoded paths, no secrets, no PII patterns

completeness No unfinished placeholders, no template markers

Grading

  • A — All checks pass, zero warnings

  • B — All checks pass, some warnings

  • C — 1–2 failures

  • D — 3+ failures

  • F — SKILL.md missing or no valid frontmatter

Analysis Dimensions (Improvement Suggestions)

Dimension What to evaluate

Description quality Length, trigger phrases, third-person voice, specificity

Workflow clarity Numbered steps, decision points, AskUserQuestion usage

Runtime robustness Preflight completeness, setup separation, degradation handling

Script quality JSON output, error handling, token awareness, exit codes

Documentation Troubleshooting tables, reference organization, no TODOs

Security Credential handling, no hardcoded paths or secrets

User experience Profile flags: has_checklist , has_completion_report , has_input_adaptation , has_language_section , has_cross_skill_handling , has_preference_persistence — for each false flag, check the "Applies when" condition in improvement_patterns.md before suggesting. A false flag with no applicable condition is the expected state, not a problem.

References

For the rationale behind each validation check, read references/validation_rules.md .

For the full knowledge base of improvement patterns with examples, read references/improvement_patterns.md .

For skill design conventions and quick reference, read references/best_practices.md .

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

General

better-skill-creator

No summary provided by upstream source.

Repository SourceNeeds Review
General

better-skill-review

No summary provided by upstream source.

Repository SourceNeeds Review
General

better-skill-publish

No summary provided by upstream source.

Repository SourceNeeds Review
General

skill-creator

No summary provided by upstream source.

Repository SourceNeeds Review