devtu-optimize-skills

Optimize ToolUniverse skills for better report quality, evidence handling, and user experience. Apply patterns like tool verification, foundation data layers, disambiguation-first, evidence grading, quantified completeness, and report-only output. Use when reviewing skills, improving existing skills, or creating new ToolUniverse research skills.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "devtu-optimize-skills" with this command: npx skills add mims-harvard/tooluniverse/mims-harvard-tooluniverse-devtu-optimize-skills

Optimizing ToolUniverse Skills

Best practices for high-quality research skills with evidence grading and source attribution.

Tool Quality Standards

  1. Error messages must be actionable — tell the user what went wrong AND what to do
  2. Schema must match API reality — run python3 -m tooluniverse.cli run <Tool> '<json>' to verify
  3. Coverage transparency — state what data is NOT included
  4. Input validation before API calls — don't silently send invalid values
  5. Cross-tool routing — name the correct tool when query is out-of-scope
  6. No silent parameter dropping — if a parameter is ignored, say so

Core Principles (13 Patterns)

Full details: references/optimization-patterns.md

#PatternKey Idea
1Tool Interface Verificationget_tool_info() before first call; maintain corrections table
2Foundation Data LayerQuery aggregator (Open Targets, PubChem) FIRST
3Versioned IdentifiersCapture both ENSG00000123456 and .12 version
4Disambiguation FirstResolve IDs, detect collisions, build negative filters
5Report-Only OutputNarrative in report; methodology in appendix only if asked
6Evidence GradingT1 (mechanistic) → T2 (functional) → T3 (association) → T4 (mention)
7Quantified CompletenessNumeric minimums per section (>=20 PPIs, top 10 tissues)
8Mandatory ChecklistAll sections exist, even if "Limited evidence"
9Aggregated Data GapsSingle section consolidating all missing data
10Query StrategyHigh-precision seeds → citation expansion → collision-filtered broad
11Tool Failure HandlingPrimary → Fallback 1 → Fallback 2 → document unavailable
12Scalable OutputNarrative report + JSON/CSV bibliography
13Synthesis SectionsBiological model + testable hypotheses, not just paper lists

Optimized Skill Workflow

Phase -1: Tool Verification (check params)
Phase  0: Foundation Data (aggregator query)
Phase  1: Disambiguation (IDs, collisions, baseline)
Phase  2: Specialized Queries (fill gaps)
Phase  3: Report Synthesis (evidence-graded narrative)

Testing Standards

Full details: references/testing-standards.md

Critical rule: NEVER write skill docs without testing all tool calls first.

  • 30+ tests per skill, 100% pass rate
  • All tests use real data (no placeholders)
  • Phase + integration + edge case tests
  • SOAP tools (IMGT, SAbDab, TheraSAbDab) need operation parameter
  • Distinguish transient errors (retry) from real bugs (fix)
  • API docs are often wrong — always verify with actual calls

Common Anti-Patterns

Anti-PatternFix
"Search Log" reportsKeep methodology internal; report findings only
Missing disambiguationAdd collision detection; build negative filters
No evidence gradingApply T1-T4 grades; label each claim
Empty sections omittedInclude with "None identified"
No synthesisAdd biological model + hypotheses
Silent failuresDocument in Data Gaps; implement fallbacks
Wrong tool parametersVerify via get_tool_info() before calling
GTEx returns nothingTry versioned ID ENSG*.version
No foundation layerQuery aggregator first
Untested tool callsTest-driven: test script FIRST

Quick Fixes for User Complaints

ComplaintFix
"Report too short"Add Phase 0 foundation + Phase 1 disambiguation
"Too much noise"Add collision filtering
"Can't tell what's important"Add T1-T4 evidence tiers
"Missing sections"Add mandatory checklist with minimums
"Too long/unreadable"Separate narrative from JSON
"Just a list of papers"Add synthesis sections
"Tool failed, no data"Add retry + fallback chains

Skill Template

---
name: [domain]-research
description: [What + when triggers]
---

# [Domain] Research

## Workflow
Phase -1: Tool Verification → Phase 0: Foundation → Phase 1: Disambiguate
→ Phase 2: Search → Phase 3: Report

## Phase -1: Tool Verification
[Parameter corrections table]

## Phase 0: Foundation Data
[Aggregator query]

## Phase 1: Disambiguation
[IDs, collisions, baseline]

## Phase 2: Specialized Queries
[Query strategy, fallbacks]

## Phase 3: Report Synthesis
[Evidence grading, mandatory sections]

## Output Files
- [topic]_report.md, [topic]_bibliography.json

## Quantified Minimums
[Numbers per section]

## Completeness Checklist
[Required sections with checkboxes]

Additional References

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Coding

devtu-create-tool

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

devtu-optimize-descriptions

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

tooluniverse-clinical-trial-design

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

devtu-auto-discover-apis

No summary provided by upstream source.

Repository SourceNeeds Review