teamwork

Dynamically creates and manages AI agent teams for complex tasks. Invoke when user requests multi-agent collaboration, complex project execution, or when tasks require specialized roles and coordinated workflow.

Safety Notice

This listing is from the official public ClawHub registry. Review SKILL.md and referenced scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "teamwork" with this command: npx skills add ChenXinBest/teamwork

Teamwork Skill

This skill enables dynamic team creation and management for executing complex engineering tasks through coordinated AI agents with intelligent model selection, cost optimization, and continuous performance evaluation.

When to Invoke

Invoke this skill when:

  • User requests execution of complex projects requiring multiple specialized roles
  • Tasks need to be broken down into coordinated steps (analysis, design, implementation, testing, review)
  • User wants to leverage multiple AI models/providers for optimal cost-performance balance
  • Projects require structured workflow with quality assurance and iteration

Initialization & Configuration Management

Automatic Initialization

IMPORTANT: This skill includes an autonomous initialization system. When invoked for the first time or when configuration is missing, it will automatically:

  1. Check Configuration Status

    • Verify if .trae/config/providers.json exists
    • Verify if .trae/config/team-roles.json exists
    • Verify if .trae/data/model_scores.json exists
  2. Interactive Setup Process If configuration files are missing or incomplete, the skill will proactively ask the user:

    Step 1: Provider Setup

    • Ask: "Which AI providers would you like to configure? (e.g., OpenAI, Anthropic, Google, Azure, etc.)"
    • For each provider, collect:
      • Provider name
      • API key (or environment variable name)
      • Base URL (if custom endpoint)

    Step 2: Model Configuration For each provider, ask:

    • "Which models from [provider] would you like to use?"
    • For each model, collect:
      • Model name/identifier
      • Pricing model type (subscription/tiered_usage/pay_per_use)
      • Pricing details based on type:
        • Subscription: cost, start date, end date
        • Tiered Usage: daily quota, monthly quota, overage rate
        • Pay-Per-Use: input cost per 1k tokens, output cost per 1k tokens
      • Capabilities (e.g., reasoning, coding, fast-response)
      • Maximum concurrent tasks

    Step 3: Host Model Selection

    • Ask: "Which model should serve as the primary interface (host model)?"
    • Present list of configured models
    • User selects one as the main interaction point

    Step 4: Budget Configuration

    • Ask: "What is your monthly budget limit? (optional)"
    • Set alert thresholds
  3. Configuration Persistence

    • Save all configurations to .trae/config/providers.json
    • Create default role definitions in .trae/config/team-roles.json
    • Initialize empty scores database in .trae/data/model_scores.json
    • Confirm successful setup with user

Configuration Management Commands

Users can manage their configuration at any time using these commands:

View Configuration

User: "Show me my current provider and model configuration"

Response: Display complete configuration from .trae/config/providers.json

Add Provider

User: "Add a new provider: [provider name]"

Action: Interactive prompts for provider details, then append to configuration

Add Model

User: "Add model [model name] to provider [provider name]"

Action: Interactive prompts for model details, then add to provider's model list

Update Model Pricing

User: "Update pricing for [model name]"

Action: Ask for new pricing details and update configuration

Remove Model

User: "Remove model [model name] from provider [provider name]"

Action: Confirm and remove from configuration

Change Host Model

User: "Change the host model to [model name]"

Action: Update host_model configuration

View Model Scores

User: "Show me the performance scores for all models"

Response: Display current model capability scores from .trae/data/model_scores.json

Reset Configuration

User: "Reset all configurations to default"

Action: Confirm with user, then reinitialize

Configuration File Structure

Provider Configuration (.trae/config/providers.json)

{
  "version": "1.0",
  "last_updated": "2026-02-12T11:00:00Z",
  "providers": [
    {
      "name": "openai",
      "api_key": "${OPENAI_API_KEY}",
      "base_url": "https://api.openai.com/v1",
      "models": [
        {
          "name": "gpt-4",
          "pricing_model": "pay_per_use",
          "input_cost_per_1k": 0.03,
          "output_cost_per_1k": 0.06,
          "context_window": 128000,
          "capabilities": ["reasoning", "coding", "analysis"],
          "max_concurrent_tasks": 3
        }
      ]
    }
  ],
  "host_model": {
    "provider": "openai",
    "model": "gpt-4"
  },
  "budget": {
    "max_monthly_cost": 100.00,
    "currency": "USD",
    "alert_threshold": 0.8
  }
}

Team Roles Configuration (.trae/config/team-roles.json)

{
  "version": "1.0",
  "last_updated": "2026-02-12T11:00:00Z",
  "roles": {
    "project_manager": {
      "description": "Coordinates team activities and manages timeline",
      "required_capabilities": ["planning", "coordination", "communication"],
      "preferred_model_traits": {
        "reliability": "high",
        "thinking_depth": "medium",
        "response_speed": "medium"
      }
    }
  }
}

Model Scores Database (.trae/data/model_scores.json)

{
  "version": "1.0",
  "last_updated": "2026-02-12T11:00:00Z",
  "evaluation_interval": 3600,
  "scores": {}
}

Initialization Checklist

Before executing any team task, verify:

  • .trae/config/providers.json exists and contains at least one provider
  • At least one model is configured
  • Host model is designated
  • .trae/config/team-roles.json exists with role definitions
  • .trae/data/model_scores.json exists (can be empty initially)

If any checklist item fails, trigger interactive initialization.

Core System Components

1. Model Performance Evaluation System

Multi-Dimensional Scoring

All models are periodically evaluated by peer models across multiple dimensions:

Evaluation Dimensions:

  • Response Speed: How quickly the model responds to requests
  • Response Frequency: Rate of successful responses within time windows
  • Thinking Depth: Quality of reasoning and problem-solving approach
  • Multi-threading Capability: Ability to handle parallel tasks
  • Code Quality: Quality of generated code (for coding tasks)
  • Creativity: Novelty and innovation in solutions
  • Reliability: Consistency in performance across sessions
  • Context Understanding: Ability to maintain context over long conversations

Scoring Mechanism:

  • Each model scores other models on a scale (e.g., 1-10) for each dimension
  • Scores are aggregated using weighted average
  • Evaluations occur after each task completion
  • Historical scores are maintained with decay factor for recent performance
  • Final capability score = weighted sum of all dimension scores

Score Storage:

{
  "model_scores": {
    "gpt-4": {
      "response_speed": 8.5,
      "response_frequency": 9.0,
      "thinking_depth": 9.5,
      "multi_threading": 7.0,
      "code_quality": 9.0,
      "creativity": 8.5,
      "reliability": 9.5,
      "context_understanding": 9.0,
      "overall_score": 8.75,
      "evaluation_count": 42,
      "last_updated": "2026-02-12T10:30:00Z"
    }
  }
}

2. Cost Calculation System

Pricing Models

Subscription-Based (订阅制)

  • Fixed cost for unlimited usage during subscription period
  • Lowest effective cost per request when fully utilized
  • Marked as expired after subscription ends → excluded from team
  • Configuration:
{
  "pricing_model": "subscription",
  "cost": 20.00,
  "currency": "USD",
  "valid_from": "2026-02-01",
  "valid_until": "2026-03-01",
  "status": "active"
}

Tiered Usage (阶段用量制)

  • Lower cost with daily/monthly quotas
  • Medium cost effectiveness
  • Must monitor quota usage daily
  • Configuration:
{
  "pricing_model": "tiered_usage",
  "daily_quota": 1000,
  "daily_used": 450,
  "monthly_quota": 30000,
  "monthly_used": 12500,
  "cost": 15.00,
  "currency": "USD",
  "overage_rate": 0.02
}

Pay-Per-Use (用量计费制)

  • Highest cost per request
  • No quota limits
  • Best for sporadic or overflow usage
  • Configuration:
{
  "pricing_model": "pay_per_use",
  "input_cost_per_1k": 0.03,
  "output_cost_per_1k": 0.06,
  "currency": "USD",
  "total_spent": 2.45
}

Cost Score Calculation:

cost_score = (normalized_cost) * (usage_efficiency) * (availability_factor)

3. Model Availability Management

Status Tracking:

  • available: Ready to accept tasks
  • busy: Currently processing tasks
  • expired: Subscription expired
  • quota_exceeded: Daily/monthly quota reached
  • rate_limited: Temporarily unavailable due to rate limits
  • offline: Provider API unavailable

Busy State Management:

{
  "model_status": {
    "gpt-4": {
      "status": "busy",
      "current_tasks": ["task-123", "task-456"],
      "max_concurrent": 3,
      "estimated_free_at": "2026-02-12T11:00:00Z"
    }
  }
}

Task Execution Workflow

Phase 1: User Request & Requirement Analysis

Step 1.1: User submits request to Host Model (主模型)

  • User interacts with designated host model (primary interface)
  • Host model receives and acknowledges the request

Step 1.2: Host model decomposes requirements

  • Break down request into phases and subtasks
  • Identify dependencies between tasks
  • Estimate complexity and required capabilities
  • Create initial task tree

Step 1.3: User confirmation

  • Present task breakdown to user
  • Clarify ambiguities and refine requirements
  • Get explicit approval to proceed

Output:

{
  "task_id": "task-789",
  "phases": [
    {
      "phase_id": "phase-1",
      "name": "Requirement Analysis",
      "subtasks": [
        {
          "subtask_id": "st-1",
          "description": "Analyze user requirements",
          "required_capabilities": ["analysis", "communication"],
          "estimated_complexity": "medium"
        }
      ]
    }
  ]
}

Phase 2: Team Assembly Meeting

Step 2.1: Host model convenes all available models

  • Filter models by status (exclude expired, offline, quota_exceeded)
  • Check busy models for potential availability
  • Send meeting invitation to all eligible models

Step 2.2: Task briefing

  • Host model presents all task content to all models
  • Share task breakdown, requirements, and constraints
  • Distribute context and background information

Step 2.3: Collaborative role definition

  • All models discuss and agree on required roles
  • Define capability requirements for each role
  • Estimate workload for each role
  • Identify potential bottlenecks

Meeting Output:

{
  "meeting_id": "meeting-456",
  "required_roles": [
    {
      "role_name": "architect",
      "required_capabilities": ["system-design", "architecture"],
      "estimated_workload": "high",
      "priority": "critical"
    },
    {
      "role_name": "developer",
      "required_capabilities": ["coding", "debugging"],
      "estimated_workload": "high",
      "priority": "high"
    }
  ],
  "consensus_reached": true
}

Phase 3: Role Assignment

Step 3.1: Self-nomination

  • Each model evaluates own suitability based on:
    • Current cost score
    • Current capability score
    • Current workload (busy status)
    • Role requirements match
  • Models submit role preferences

Step 3.2: Conflict resolution

  • If multiple models want same role: democratic voting
  • If role has no candidates: negotiate with best-fit models
  • Balance workload distribution across models
  • Avoid concentrating all tasks on single model

Step 3.3: Final assignment

  • Confirm role assignments with all models
  • Document assignment rationale
  • Update model busy status

Assignment Algorithm:

for each role:
  candidates = models.filter(capable_and_available)
  if len(candidates) == 1:
    assign to candidates[0]
  elif len(candidates) > 1:
    scores = calculate_combined_score(candidates, role)
    winner = vote_among_models(candidates, scores)
    assign to winner
  else:
    negotiate_with_best_available_model()

Combined Score Calculation:

combined_score = (capability_score * 0.4) + 
                 (cost_efficiency_score * 0.3) + 
                 (availability_score * 0.2) + 
                 (workload_balance_factor * 0.1)

Phase 4: Herald Selection & Communication Setup

Step 4.1: Select Herald (传令官)

  • Choose fastest responding model (not necessarily most capable)
  • Herald acts as central communication hub
  • All models communicate through herald

Herald Responsibilities:

  • Relay messages between all team members
  • Distribute progress updates
  • Broadcast requirements and instructions
  • Collect and aggregate results
  • Monitor task completion status
  • Report status to host model
  • Handle timeout and failure notifications

Step 4.2: Communication channels

Model A → Herald → Model B
Model A → Herald → All Models
Herald → Host Model (status reports)

Herald Configuration:

{
  "herald": {
    "model": "gpt-3.5-turbo",
    "selection_criteria": "fastest_response",
    "polling_interval": 30,
    "timeout_threshold": 300,
    "responsibilities": [
      "message_relay",
      "status_monitoring",
      "progress_tracking",
      "failure_reporting"
    ]
  }
}

Phase 5: Task Execution

Step 5.1: Parallel execution

  • Assigned models work on their respective tasks
  • Regular progress updates to herald
  • Herald broadcasts relevant updates to team

Step 5.2: Coordination

  • Herald checks task status periodically
  • Identifies blockers and delays
  • Facilitates inter-model communication
  • Escalates issues to host model

Step 5.3: Progress tracking

{
  "task_progress": {
    "task_id": "task-789",
    "overall_progress": 65,
    "subtask_status": {
      "st-1": "completed",
      "st-2": "in_progress",
      "st-3": "pending"
    },
    "blockers": [],
    "estimated_completion": "2026-02-12T14:00:00Z"
  }
}

Phase 6: Task Completion & Review Meeting

Step 6.1: Completion notification

  • Herald confirms all tasks completed
  • Collects final outputs from all models
  • Aggregates results

Step 6.2: Summary meeting

  • Host model convenes all participating models
  • Each model presents their contribution
  • Discuss challenges and solutions
  • Evaluate collaboration effectiveness

Step 6.3: Performance re-evaluation

  • Models rate each other's performance
  • Update capability scores based on task execution
  • Record role-model fit assessments
  • Update model scores database

Evaluation Form:

{
  "evaluation": {
    "evaluator": "gpt-4",
    "evaluatee": "claude-3",
    "task_id": "task-789",
    "role_played": "developer",
    "scores": {
      "response_speed": 8,
      "thinking_depth": 9,
      "code_quality": 9,
      "collaboration": 8
    },
    "role_fit": "excellent",
    "comments": "Strong problem-solving skills"
  }
}

Phase 7: Failure Handling & Iteration

Step 7.1: Failure detection

  • Herald detects task failure or timeout
  • Collects failure information from relevant models
  • Reports to host model with detailed context

Step 7.2: Failure analysis meeting

  • Convene all participating models
  • Analyze root cause of failure
  • Identify contributing factors
  • Propose solutions

Step 7.3: User consultation

  • Host model presents failure analysis to user
  • Discuss potential solutions:
    • Requirement changes
    • Approach modifications
    • Team reconfiguration
    • Additional resources
  • Get user decision on next steps

Step 7.4: Iteration or termination

  • If user approves changes: restart from appropriate phase
  • If user terminates: document lessons learned
  • Update model scores based on partial performance

Configuration Files

Provider Configuration (.trae/config/providers.json)

{
  "providers": [
    {
      "name": "openai",
      "api_key": "${OPENAI_API_KEY}",
      "base_url": "https://api.openai.com/v1",
      "models": [
        {
          "name": "gpt-4",
          "pricing_model": "pay_per_use",
          "input_cost_per_1k": 0.03,
          "output_cost_per_1k": 0.06,
          "context_window": 128000,
          "capabilities": ["reasoning", "coding", "analysis"],
          "max_concurrent_tasks": 3
        },
        {
          "name": "gpt-3.5-turbo",
          "pricing_model": "subscription",
          "subscription_cost": 20.00,
          "valid_from": "2026-02-01",
          "valid_until": "2026-03-01",
          "context_window": 16385,
          "capabilities": ["fast-response", "coding"],
          "max_concurrent_tasks": 5
        }
      ]
    },
    {
      "name": "anthropic",
      "api_key": "${ANTHROPIC_API_KEY}",
      "base_url": "https://api.anthropic.com",
      "models": [
        {
          "name": "claude-3-opus",
          "pricing_model": "tiered_usage",
          "daily_quota": 500,
          "monthly_quota": 15000,
          "input_cost_per_1k": 0.015,
          "output_cost_per_1k": 0.075,
          "context_window": 200000,
          "capabilities": ["reasoning", "analysis", "long-context"],
          "max_concurrent_tasks": 2
        }
      ]
    }
  ],
  "host_model": {
    "provider": "openai",
    "model": "gpt-4",
    "role": "primary_interface"
  },
  "budget": {
    "max_monthly_cost": 100.00,
    "currency": "USD",
    "alert_threshold": 0.8
  }
}

Team Roles Configuration (.trae/config/team-roles.json)

{
  "roles": {
    "project_manager": {
      "description": "Coordinates team activities and manages timeline",
      "required_capabilities": ["planning", "coordination", "communication"],
      "preferred_model_traits": {
        "reliability": "high",
        "thinking_depth": "medium",
        "response_speed": "medium"
      },
      "typical_workload": "medium"
    },
    "architect": {
      "description": "Designs system architecture and technical approach",
      "required_capabilities": ["system-design", "architecture", "patterns"],
      "preferred_model_traits": {
        "thinking_depth": "high",
        "creativity": "high",
        "context_understanding": "high"
      },
      "typical_workload": "high"
    },
    "developer": {
      "description": "Implements code following specifications",
      "required_capabilities": ["coding", "debugging", "refactoring"],
      "preferred_model_traits": {
        "code_quality": "high",
        "response_speed": "medium",
        "reliability": "high"
      },
      "typical_workload": "high"
    },
    "tester": {
      "description": "Creates and executes test suites",
      "required_capabilities": ["testing", "qa", "validation"],
      "preferred_model_traits": {
        "thinking_depth": "medium",
        "response_speed": "high",
        "reliability": "high"
      },
      "typical_workload": "medium"
    },
    "reviewer": {
      "description": "Performs code reviews and quality checks",
      "required_capabilities": ["code-review", "best-practices", "security"],
      "preferred_model_traits": {
        "thinking_depth": "high",
        "code_quality": "high",
        "reliability": "high"
      },
      "typical_workload": "medium"
    },
    "analyst": {
      "description": "Analyzes requirements and breaks down tasks",
      "required_capabilities": ["analysis", "communication", "documentation"],
      "preferred_model_traits": {
        "thinking_depth": "high",
        "context_understanding": "high",
        "creativity": "medium"
      },
      "typical_workload": "medium"
    }
  }
}

Model Scores Database (.trae/data/model_scores.json)

{
  "last_evaluation": "2026-02-12T10:30:00Z",
  "evaluation_interval": 3600,
  "scores": {
    "gpt-4": {
      "dimensions": {
        "response_speed": 8.5,
        "response_frequency": 9.0,
        "thinking_depth": 9.5,
        "multi_threading": 7.0,
        "code_quality": 9.0,
        "creativity": 8.5,
        "reliability": 9.5,
        "context_understanding": 9.0
      },
      "overall_score": 8.75,
      "evaluation_count": 42,
      "role_fit_history": {
        "architect": 9.2,
        "developer": 8.8,
        "reviewer": 9.0
      }
    }
  }
}

Best Practices

Model Selection

  • Match model capabilities to task requirements
  • Consider cost-effectiveness for routine tasks
  • Reserve high-capability models for complex reasoning
  • Distribute workload to prevent bottlenecks

Communication

  • Keep messages concise and clear
  • Use structured formats for inter-model communication
  • Herald should batch non-urgent updates
  • Escalate critical issues immediately

Performance Optimization

  • Cache frequently used context
  • Batch similar requests when possible
  • Monitor quota usage proactively
  • Maintain backup models for critical roles

Quality Assurance

  • Always conduct review meetings
  • Update scores after each task
  • Learn from failures systematically
  • Continuously refine role definitions

Error Handling

Provider Failures

  • Retry with exponential backoff
  • Switch to backup provider
  • Notify team of delays
  • Update model availability status

Task Failures

  • Capture detailed error context
  • Analyze root cause with team
  • Propose remediation strategies
  • Consult user for major changes

Communication Failures

  • Herald implements heartbeat checks
  • Fallback to direct model-to-model communication
  • Reassign herald if unresponsive
  • Log all communication issues

Output Format

Final deliverables include:

  • Complete task execution report
  • Team composition and role assignments
  • Individual model performance metrics
  • Cost breakdown and usage statistics
  • Updated model capability scores
  • Lessons learned and recommendations

Skill Structure & Components

Directory Structure

.trae/skills/teamwork/
├── SKILL.md                    # Main skill definition (this file)
├── scripts/                    # Execution scripts
│   ├── init.js                # Initialization and configuration loader
│   ├── config-manager.js      # Provider and model configuration management
│   ├── score-manager.js       # Model performance score management
│   ├── team-coordinator.js    # Team assembly and task coordination
│   └── herald.js              # Communication and message relay system
├── templates/                  # Document templates
│   ├── task-report.md         # Task execution report template
│   ├── meeting-minutes.md     # Meeting minutes template
│   ├── failure-report.md      # Failure analysis report template
│   └── evaluation-form.md     # Model evaluation form template
├── utils/                      # Utility functions
│   ├── index.js               # Utility exports
│   ├── helpers.js             # General helper functions
│   ├── logger.js              # Logging system
│   ├── template-renderer.js   # Template rendering engine
│   └── errors.js              # Custom error classes
└── data/                       # Skill runtime data
    └── (generated at runtime)

Scripts Reference

init.js - Initialization Module

Purpose: Handles skill initialization and configuration loading.

Key Functions:

  • ensureDirectories() - Create required directories
  • checkConfiguration() - Verify configuration status
  • initializeDefaultRoles() - Create default role definitions
  • initializeEmptyScores() - Initialize empty scores database
  • initializeEmptyProviders() - Initialize empty providers config
  • needsInitialization() - Check if initialization is required
  • readJSON(filePath) - Read JSON configuration file
  • writeJSON(filePath, data) - Write JSON configuration file

Usage:

const init = require('./scripts/init.js');

// Check if initialization needed
if (init.needsInitialization()) {
  init.initializeDefaultRoles();
  init.initializeEmptyScores();
  init.initializeEmptyProviders();
}

config-manager.js - Configuration Management

Purpose: Manage provider and model configurations.

Key Functions:

  • addProvider(config, providerInfo) - Add new provider
  • addModel(config, providerName, modelInfo) - Add model to provider
  • removeModel(config, providerName, modelName) - Remove model
  • removeProvider(config, providerName) - Remove provider
  • updateModelPricing(config, providerName, modelName, pricingInfo) - Update pricing
  • setHostModel(config, providerName, modelName) - Set host model
  • setBudget(config, budgetInfo) - Set budget limits
  • getAvailableModels(config) - Get list of available models
  • getModelStatus(model) - Get model availability status
  • displayConfiguration(config) - Display current configuration

Usage:

const configManager = require('./scripts/config-manager.js');
const config = init.readJSON(init.PROVIDERS_FILE);

// Add new provider
configManager.addProvider(config, {
  name: 'openai',
  api_key: '${OPENAI_API_KEY}',
  base_url: 'https://api.openai.com/v1'
});

// Add model with subscription pricing
configManager.addModel(config, 'openai', {
  name: 'gpt-4',
  pricing_model: 'subscription',
  subscription_cost: 20.00,
  valid_from: '2026-02-01',
  valid_until: '2026-03-01',
  capabilities: ['reasoning', 'coding']
});

score-manager.js - Performance Score Management

Purpose: Manage model performance evaluation scores.

Key Functions:

  • initializeModelScore(scores, modelName, provider) - Initialize model scores
  • updateModelScore(scores, modelName, dimension, newScore, evaluator) - Update dimension score
  • calculateOverallScore(dimensions, weights) - Calculate weighted overall score
  • updateRoleFit(scores, modelName, roleName, fitScore) - Update role fit score
  • getTopModelsForRole(scores, roleName, topN) - Get top models for a role
  • getModelsByCapability(scores, capability, minScore) - Get models by capability
  • recordEvaluation(scores, evaluation) - Record complete evaluation
  • displayScores(scores) - Display all model scores

Usage:

const scoreManager = require('./scripts/score-manager.js');

// Record evaluation
scoreManager.recordEvaluation(scores, {
  evaluator: 'gpt-4',
  evaluatee: 'claude-3',
  task_id: 'task-123',
  role_played: 'developer',
  scores: {
    response_speed: 8,
    thinking_depth: 9,
    code_quality: 9
  },
  role_fit: 'excellent'
});

// Get top models for architect role
const topArchitects = scoreManager.getTopModelsForRole(scores, 'architect', 3);

team-coordinator.js - Team Coordination

Purpose: Coordinate team assembly and task execution.

Key Class: TeamCoordinator

Methods:

  • load() - Load configurations
  • getAvailableModels() - Get available models list
  • selectHerald() - Select fastest model as herald
  • assignRoles(requiredRoles) - Assign roles to models
  • calculateCombinedScore(model, roleFit) - Calculate selection score
  • createTaskPlan(userRequest) - Create task execution plan
  • generateMeetingAgenda(meetingType) - Generate meeting agenda
  • generateEvaluationForms() - Generate peer evaluation forms
  • generateReport() - Generate task report

Usage:

const TeamCoordinator = require('./scripts/team-coordinator.js');

const coordinator = new TeamCoordinator();
coordinator.load();

// Select herald
const herald = coordinator.selectHerald();

// Assign roles
const assignments = coordinator.assignRoles(['architect', 'developer', 'tester']);

// Create task plan
const plan = coordinator.createTaskPlan('Build a REST API');

herald.js - Communication System

Purpose: Manage inter-model communication and coordination.

Key Class: Herald

Methods:

  • initializeTeam(team) - Initialize team status tracking
  • broadcast(message, excludeSender) - Broadcast message to all
  • sendDirectMessage(to, message, from) - Send direct message
  • updateProgress(model, subtaskId, progress, status) - Update task progress
  • getTeamStatus() - Get current team status
  • checkTimeouts() - Check for timeout conditions
  • pollTeam() - Request status from all members
  • reportToHost(status) - Send status report to host
  • notifyFailure(model, error, context) - Notify failure
  • notifyCompletion(model, result) - Notify completion
  • getOverallProgress() - Get overall task progress

Usage:

const Herald = require('./scripts/herald.js');

const herald = new Herald('gpt-3.5-turbo', 'openai');
herald.initializeTeam(team);

// Broadcast update
herald.broadcast({ type: 'task_update', content: 'Phase 1 complete' });

// Check progress
const progress = herald.getOverallProgress();

Templates Reference

task-report.md - Task Execution Report

Purpose: Document complete task execution details.

Variables:

  • task_id - Unique task identifier
  • timestamp - Report generation time
  • status - Task status (completed/failed/in_progress)
  • summary - Executive summary
  • team_members - Array of team member details
  • phases - Array of execution phases
  • model_metrics - Performance metrics per model
  • total_cost - Total execution cost
  • deliverables - Array of deliverables
  • lessons - Lessons learned
  • recommendations - Recommendations
  • score_updates - Model score updates

Usage:

const { renderTemplateFromFile } = require('./utils/template-renderer.js');

const report = renderTemplateFromFile('task-report.md', {
  task_id: 'task-123',
  timestamp: new Date().toISOString(),
  status: 'completed',
  summary: 'Successfully implemented REST API',
  team_members: [...],
  phases: [...]
});

meeting-minutes.md - Meeting Documentation

Purpose: Document team meeting discussions and decisions.

Variables:

  • meeting_id - Unique meeting identifier
  • meeting_type - Type of meeting
  • date - Meeting date
  • duration - Meeting duration
  • participants - Array of participants
  • agenda_items - Meeting agenda
  • voting_results - Voting results (if applicable)
  • action_items - Action items from meeting
  • next_steps - Next steps to take

failure-report.md - Failure Analysis

Purpose: Document and analyze task failures.

Variables:

  • task_id - Failed task identifier
  • failure_time - Time of failure
  • failure_type - Type of failure
  • severity - Failure severity
  • timeline - Timeline of events
  • primary_cause - Root cause
  • contributing_factors - Contributing factors
  • recovery_actions - Actions taken for recovery
  • recommendations - Recommendations to prevent recurrence

evaluation-form.md - Model Evaluation

Purpose: Document peer model evaluations.

Variables:

  • evaluator_model - Evaluating model
  • evaluatee_model - Model being evaluated
  • task_id - Related task
  • role_played - Role in task
  • response_speed through context_understanding - Dimension scores
  • role_fit - Overall role fit assessment
  • strengths - Model strengths
  • improvements - Areas for improvement

Utilities Reference

helpers.js - General Utilities

Functions:

  • generateId(prefix) - Generate unique identifier
  • formatDate(date) - Format date to ISO string
  • formatDuration(ms) - Format milliseconds to readable duration
  • calculateCost(model, inputTokens, outputTokens) - Calculate API cost
  • deepClone(obj) - Deep clone object
  • mergeObjects(target, source) - Deep merge objects
  • retryWithBackoff(fn, maxRetries, delay) - Retry with exponential backoff
  • chunkArray(array, size) - Split array into chunks
  • groupBy(array, key) - Group array by key
  • sortBy(array, key, order) - Sort array by key
  • uniqueBy(array, key) - Remove duplicates by key

logger.js - Logging System

Classes: Logger

Log Levels: DEBUG, INFO, WARN, ERROR

Methods:

  • debug(message, data) - Log debug message
  • info(message, data) - Log info message
  • warn(message, data) - Log warning message
  • error(message, data) - Log error message
  • setLevel(level) - Set log level
  • setLogFile(filePath) - Set log file path

Usage:

const { createLogger, LOG_LEVELS } = require('./utils/logger.js');

const logger = createLogger('teamwork', { 
  level: LOG_LEVELS.DEBUG,
  console: true 
});

logger.info('Task started', { task_id: 'task-123' });

errors.js - Custom Errors

Error Classes:

  • ValidationError - Input validation errors
  • ConfigurationError - Configuration errors
  • ModelNotFoundError - Model not found errors
  • ProviderNotFoundError - Provider not found errors
  • TaskExecutionError - Task execution errors
  • TimeoutError - Timeout errors
  • BudgetExceededError - Budget exceeded errors
  • QuotaExceededError - Quota exceeded errors
  • HeraldError - Herald communication errors

Functions:

  • handleError(error, logger) - Standardized error handling
  • isRecoverable(error) - Check if error is recoverable

API Reference

Quick Start

// 1. Initialize skill
const init = require('./scripts/init.js');
if (init.needsInitialization()) {
  // Run interactive setup
  init.initializeDefaultRoles();
  init.initializeEmptyScores();
  init.initializeEmptyProviders();
}

// 2. Configure providers
const configManager = require('./scripts/config-manager.js');
const config = init.readJSON(init.PROVIDERS_FILE);

configManager.addProvider(config, { name: 'openai' });
configManager.addModel(config, 'openai', {
  name: 'gpt-4',
  pricing_model: 'pay_per_use',
  input_cost_per_1k: 0.03,
  output_cost_per_1k: 0.06,
  capabilities: ['reasoning', 'coding']
});
configManager.setHostModel(config, 'openai', 'gpt-4');
init.writeJSON(init.PROVIDERS_FILE, config);

// 3. Create team and execute task
const TeamCoordinator = require('./scripts/team-coordinator.js');
const coordinator = new TeamCoordinator();
coordinator.load();

const herald = coordinator.selectHerald();
const team = coordinator.assignRoles(['architect', 'developer', 'tester']);
const plan = coordinator.createTaskPlan('Build REST API');

// 4. Execute with herald coordination
const Herald = require('./scripts/herald.js');
const heraldInstance = new Herald(herald.model, herald.provider);
heraldInstance.initializeTeam(team);

// 5. Record evaluations and update scores
const scoreManager = require('./scripts/score-manager.js');
const scores = init.readJSON(init.SCORES_FILE);

scoreManager.recordEvaluation(scores, {
  evaluator: 'gpt-4',
  evaluatee: 'claude-3',
  task_id: plan.task_id,
  role_played: 'developer',
  scores: { response_speed: 8, thinking_depth: 9, code_quality: 9 },
  role_fit: 'excellent'
});

init.writeJSON(init.SCORES_FILE, scores);

Version History

  • v1.0.0 (2026-02-12): Initial release with full feature set
    • Multi-provider support
    • Three pricing models
    • 8-dimension performance evaluation
    • Herald communication system
    • Complete workflow management
    • Template-based reporting

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Automation

Ai Automation Consulting

AI 自动化咨询服务 - 帮你用 AI 省时省钱。适合:中小企业、自由职业者、想提效的人。

Registry SourceRecently Updated
Automation

myskill

Helps users discover and install agent skills when they ask questions like "how do I do X", "find a skill for X", "is there a skill that can...", or express...

Registry SourceRecently Updated
Automation

GridClash

Battle in Grid Clash - join 8-agent grid battles. Fetch equipment data to choose the best weapon, armor, and tier. Use when user wants to participate in Grid...

Registry SourceRecently Updated