analyze-results

Analyze ML experiment results, compute statistics, generate comparison tables and insights. Use when user says "analyze results", "compare", or needs to interpret experimental data.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "analyze-results" with this command: npx skills add wanshuiyin/auto-claude-code-research-in-sleep/wanshuiyin-auto-claude-code-research-in-sleep-analyze-results

Analyze Experiment Results

Analyze: $ARGUMENTS

Workflow

Step 1: Locate Results

Find all relevant JSON/CSV result files:

  • Check figures/, results/, or project-specific output directories
  • Parse JSON results into structured data

Step 2: Build Comparison Table

Organize results by:

  • Independent variables: model type, hyperparameters, data config
  • Dependent variables: primary metric (e.g., perplexity, accuracy, loss), secondary metrics
  • Delta vs baseline: always compute relative improvement

Step 3: Statistical Analysis

  • If multiple seeds: report mean +/- std, check reproducibility
  • If sweeping a parameter: identify trends (monotonic, U-shaped, plateau)
  • Flag outliers or suspicious results

Step 4: Generate Insights

For each finding, structure as:

  1. Observation: what the data shows (with numbers)
  2. Interpretation: why this might be happening
  3. Implication: what this means for the research question
  4. Next step: what experiment would test the interpretation

Step 5: Update Documentation

If findings are significant:

  • Propose updates to project notes or experiment reports
  • Draft a concise finding statement (1-2 sentences)

Output Format

Always include:

  1. Raw data table
  2. Key findings (numbered, concise)
  3. Suggested next experiments (if any)

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Coding

idea-creator

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

run-experiment

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

auto-review-loop

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

research-lit

No summary provided by upstream source.

Repository SourceNeeds Review
analyze-results | V50.AI