token-kill

Reduce OpenClaw token consumption by 95%+ using three optimization techniques (slash commands, script-first principle, and model tiering)

Safety Notice

This listing is from the official public ClawHub registry. Review SKILL.md and referenced scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "token-kill" with this command: npx skills add niuben/token-kill

Token Kill - OpenClaw Token Optimizer

Need help optimizing your OpenClaw token usage costs? This Skill will guide you through three powerful optimization techniques to dramatically reduce token consumption.

Based on real-world case studies, applying these optimization techniques can reduce token consumption from $200+/day to $10/day, achieving a 95%+ cost reduction.

Three Core Token Optimization Techniques

1️⃣ Slash Commands Optimization

  • /new - Start a fresh conversation and clear old context (saves 50,000+ tokens)
  • /compress - Compress memory by keeping important info and forgetting details (saves 30,000+ tokens)
  • /stop - Immediately stop current task to prevent further token consumption
  • /restart - Restart the system to clear lag and resolve issues

2️⃣ Script-First Principle

Core Philosophy: AI is your brain, not your hands

Automate with scripts instead of using the model for mechanical tasks:

  • 📧 Email Checking - Scripts monitor emails; AI only notified of new messages ($100+/month → <$1/month)
  • 🌤️ Weather Queries - Direct API calls, zero token consumption
  • 📊 Data Fetching - Scripts retrieve data; AI only handles formatting
  • Scheduled Tasks - Scripts execute; prevent AI from polling
  • 🔄 Data Processing - Script handles transformations

3️⃣ Model Tiering Strategy

Use premium models for complex tasks, budget models for simple ones

ComplexityRecommended ModelCostUse CasesSavings
🔴 HighGPT-4 / Claude$0.03/1k tokensCode generation, creative writing, complex reasoningBaseline
🟡 MediumGPT-3.5-Turbo / Ernie$0.0005/1k tokensGeneral tasks, text editing98%
🟢 LowQwen, Tongyi (Budget Models)$0.00001/1k tokensData processing, report generation, formatting99.97%

Real-World Cost Reduction Cases

Case 1: Email Monitoring System

Problem: Model checks emails every 5 minutes

ApproachMonthly Cost
❌ Model Polling$100+/month
✅ Script + AI Notification<$1/month
Savings99%

Case 2: Daily Report Generation

Scenario: Generate reports every 30 minutes (2000 tokens/call)

ModelDaily CostMonthly CostSavings
GPT-4$2.88$86Baseline
GPT-3.5$0.048$1.4498%
Qwen$0.001$0.0399.97%

Examples

Example 1: Compressing Large Memory

Scenario: After many conversations, memory.md has grown to hundreds of thousands of characters

Solution:

  1. Execute /compress command
  2. System removes trivial details while preserving core information
  3. Memory size reduced by 30-50%

Result: Reduced context loading on each turn, saves 30,000+ tokens

Example 2: Replacing AI with Scripts

Scenario: Need to check for new orders every hour

Wrong Approach:

Have model check orders API every hour
→ Model must understand and judge each time
→ 24 checks per day = huge costs

Correct Approach:

Script checks order API every hour
Notify model only on new orders
Model handles decision-making only

Savings: Script uses only CPU, saves 90%+ tokens

Example 3: Model Tiering Workflow

Scenario: Handle various complexity levels

Strategy:

  • 💻 Code Writing → GPT-4 (worth the investment)
  • 📝 Content Editing → GPT-3.5 (good balance)
  • 📊 Report Generation → Budget Model (fully sufficient)

Result: 90% cost reduction, zero functionality loss

Guidelines

✅ Best Practices for Token Savings

1. Use Slash Commands Regularly

  • Execute /compress once daily - Prevent memory bloat
  • Use /new for long conversations - Start fresh after 1+ hours
  • Use /stop on wrong tasks - Stop immediately to prevent waste

2. Strictly Follow Script-First Principle

  • Scripts handle: Scheduled checks, data fetching, API calls, data processing
  • Never let AI handle: Polling, mechanical work, repetitive checks, resource-intensive operations
  • 💡 Core rule: AI = decision-making and judgment; Scripts = execution and heavy lifting

3. Enforce Model Tiering

Task TypeModel ChoiceReason
Code generation, deep analysisGPT-4Complex tasks worth the cost
General tasks, text editingGPT-3.5Best value proposition
Data processing, reportsBudget ModelsFully capable, lowest cost

4. Regular Token Usage Audit

  • Review billing distribution
  • Identify high-cost tasks for optimization
  • Adjust model configuration and scripts

❌ Common Token Wastage Patterns

Bad PracticeConsequenceSolution
Unlimited conversation historyGrowing memory = more tokensRegular /compress or /new
AI polling for updatesToken burn on each checkUse scripts instead
Using GPT-4 for simple tasksOverkill, high costUse appropriate model tier
Never compressing memoryLinear token cost growthEstablish compression habit
Continuing failed tasksWasted tokensUse /stop immediately

Token Cost Formula

Total Cost = Context Consumption + Task Consumption

Optimization Formula:
New Cost = (Original Context × 30%) + (Task Cost × 20%)
         = Original Cost × (0.3 + 0.2)
         = Original Cost × 0.5 or lower

Combining all three techniques achieves 95%+ cost reduction.

Key Principle

💡 Remember: High costs don't come from AI itself, but from making it do tasks it shouldn't do and remember information it shouldn't store.

Assign the right tasks to the right tools, and AI becomes truly cost-effective.

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

General

Img2img

Generate images from text descriptions using DALL-E 3 while adhering to usage policies and avoiding realistic human faces.

Registry SourceRecently Updated
General

Habitat-GS-Navigator

Navigate and interact with photo-realistic 3DGS environments via the Habitat-GS Bridge. Use when: user asks to explore a 3D scene, perform embodied navigatio...

Registry SourceRecently Updated
General

Memory Palace

持久化记忆管理。Use when: 用户告诉你个人信息/偏好/习惯、需要记住项目状态/技术决策、完成任务后有可复用经验、用户说"记住""别忘了""下次注意"、需要回忆之前的对话内容。支持语义搜索和时间推理。

Registry SourceRecently Updated
General

Podcast Transcript Mining Authority Positioning

Extract guest appearances, speaking topics, and soundbites from podcast transcripts to build authority portfolios and generate podcast pitch templates. Use w...

Registry SourceRecently Updated