ClawCache Free

Smart LLM cost tracking and caching for Python

Safety Notice

This listing is from the official public ClawHub registry. Review SKILL.md and referenced scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "ClawCache Free" with this command: npx skills add abyousef739/clawcache-free

ClawCache Free - LLM Cost Tracking & Caching

ClawCache is a production-ready Python library that helps you track every penny spent on LLM APIs and automatically cache responses to slash costs.

🎯 What You Get

💰 Cost Tracking

  • Automatic logging of every LLM API call with precise token counting
  • Daily CLI reports showing spending, savings, and cache efficiency
  • Multi-provider support: OpenAI, Anthropic, Mistral, Ollama, and more
  • 2026 pricing built-in for accurate cost calculations

⚡ Smart Caching

  • Exact-match caching using SQLite (fast, reliable, local)
  • 58.3% cache hit rate proven in real-world scenarios
  • Automatic savings - cached responses cost $0
  • Composite cache keys for better accuracy (model + temperature + params)

📊 Real-World Performance

Based on comprehensive simulation with 48 API calls across 4 common use cases:

MetricValue
Cache Hit Rate58.3%
Total Cost$0.0062
API Calls Saved28 out of 48
Scenarios TestedCode Review, Data Analysis, Content Generation, QA Support

Scenario Breakdown

ScenarioCallsCache HitsHit Rate
Code Review12758.3%
Data Analysis12866.7%
Content Generation12758.3%
QA Support12650.0%

🚀 Quick Start

Installation

pip install clawcache

Basic Usage

from clawcache.free.cost import async_monitor_cost
from clawcache.free.cache_basic import BasicCache

# Initialize cache
cache = BasicCache()

# Decorate your LLM function
@async_monitor_cost
async def my_llm_call(prompt, model="gpt-4-turbo"):
    # Check cache first
    cached = await cache.aget(prompt, model=model)
    if cached:
        return cached.content
    
    # Make actual API call
    response = await openai.ChatCompletion.acreate(
        model=model,
        messages=[{"role": "user", "content": prompt}]
    )
    
    # Cache the response
    await cache.aset(prompt, response, model=model)
    return response

# Use it
result = await my_llm_call("Explain quantum computing")

View Your Cost Report

ClawCache automatically tracks all your LLM spending:

# See today's detailed cost report
clawcache --report

# Output shows:
# - Money spent today
# - Money saved via cache
# - Total API calls
# - Cache hit rate
# - Efficiency metrics

✨ Features

Cost Tracking & Monitoring

  • Automatic Cost Logging: Every API call tracked with timestamp, model, tokens, and cost
  • Daily CLI Reports: Shows spending, savings, and efficiency metrics
  • Accurate Token Counting: Uses tiktoken when available
  • Multi-Provider Support: OpenAI, Anthropic, Mistral, Ollama, etc.

Smart Caching

  • Exact-Match Caching: SQLite-based (fast and reliable)
  • Composite Cache Keys: Cache by prompt + model + params
  • Async Support: Full async/await compatibility
  • Automatic Savings: Cached responses cost $0

Security & Reliability

  • Secure: Pickle opt-in (disabled by default)
  • Concurrent-Safe: SQLite WAL mode
  • Cross-Platform: Windows, macOS, Linux

🔒 Security

ClawCache takes security seriously:

  • Pickle opt-in: Deserialization disabled by default to prevent RCE
  • SQLite WAL mode: Safe concurrent access
  • File locking: Cross-platform file locking for log integrity

📖 Configuration

Customize ClawCache behavior via environment variables:

export CLAWCACHE_HOME=/path/to/cache  # Default: ~/.clawcache

Cache Key Specificity

ClawCache supports composite cache keys for better accuracy:

# Cache by prompt + model + temperature
await cache.aset(
    prompt, 
    response, 
    model="gpt-4-turbo",
    temperature=0.7
)

Supported Models (2026 Pricing)

ModelInput ($/1M tokens)Output ($/1M tokens)
GPT-4 Turbo$10.00$30.00
GPT-3.5 Turbo$0.50$1.50
Claude 3.5 Sonnet$3.00$15.00
Claude 3 Haiku$0.25$1.25

💡 Use Cases

1. Code Review Assistant

@async_monitor_cost
async def review_code(code_snippet):
    prompt = f"Review this code for bugs: {code_snippet}"
    return await llm_call(prompt, model="gpt-4-turbo")

2. Data Analysis

@async_monitor_cost
async def analyze_data(dataset):
    prompt = f"Analyze this dataset: {dataset}"
    return await llm_call(prompt, model="claude-3-5-sonnet")

3. Content Generation

@async_monitor_cost
async def generate_content(topic):
    prompt = f"Write a blog post about: {topic}"
    return await llm_call(prompt, model="gpt-3.5-turbo")

📈 Cost Savings Projection

Based on typical usage patterns:

  • Without ClawCache: $0.0062 for 48 calls
  • With ClawCache: $0.0062 for first run, ~$0.0026 for subsequent runs (58% savings)
  • Annual Projection: For 10,000 calls/month → $3,600 saved/year

⭐ Pro Version Coming Soon

Want even more savings and insights? ClawCache Pro will include:

  • 🔮 Semantic Caching: Match similar queries (higher hit rates!)
  • 📊 Advanced Analytics: Detailed cost breakdowns and trends
  • 📈 Visual Reports: Beautiful charts and graphs
  • 🚀 Social Sharing: Share savings on Twitter, LinkedIn, Molbook with auto-generated charts
  • ☁️ Cloud Sync: Sync cache across devices
  • 🎯 Team Analytics: Track costs across your team

Free: Cost tracking with CLI reports + exact-match caching
Pro: Adds social sharing with charts + semantic caching + advanced analytics

Learn more

🤝 Contributing

Contributions welcome! Please:

  1. Fork the repository
  2. Create a feature branch
  3. Add tests for new features
  4. Submit a pull request

📄 License

MIT License - see LICENSE for details

🔗 Links


Made with ❤️ for the AI community

Save money. Track costs. Build better.

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Coding

Meaningful-Code-Context-Scanner

Graph-based code intelligence for any codebase. Scan a repo once, then query it in plain English — via browser UI or API. Returns ranked, explained results f...

Registry SourceRecently Updated
2721Profile unavailable
Automation

LLM Cost Tracker

Track OpenClaw LLM token usage and cost from OpenRouter API. Reports: last 24h, 7d, 30d, 90d, 365d with model breakdown. Skills that trigger this: - "llm cos...

Registry SourceRecently Updated
1300Profile unavailable
Coding

Slash Tokens

CLI proxy that reduces LLM token consumption by 60-90%. Prefix any dev command with 'rtk' to get filtered, compact output. Use for all Bash commands to save...

Registry SourceRecently Updated
1880Profile unavailable
Web3

Lsp Assist

Language Server Protocol integration for OpenClaw agents. Enables precise code navigation: go-to-definition, find-references, hover type info, diagnostics, a...

Registry SourceRecently Updated
2340Profile unavailable