promptcache — LoomLens Advisor
What It Does
Estimates the cost savings from caching frequently-used prompts. Compares the cost of re-sending full context every call vs. cached prompt mode across all major models.
When to Use
- When you run the same system prompt repeatedly
- Before enabling prompt caching on a production pipeline
- When evaluating cost savings from prompt template reuse
Syntax
/promptcache "You are an expert radiologist..." --calls-per-day 50
/promptcache "Summarize this in 3 bullets" --model openai/gpt-4o-mini
Free Tier
3 analyses/day free with any Signalloom API key.
Get your free key: https://signalloomai.com/signup