token-budget-advisor

Offers the user an informed choice about how much response depth to consume before answering. Use this skill when the user explicitly wants to control response length, depth, or token budget. TRIGGER when: "token budget", "token count", "token usage", "token limit", "response length", "answer depth", "short version", "brief answer", "detailed answer", "exhaustive answer", "respuesta corta vs larga", "cuántos tokens", "ahorrar tokens", "responde al 50%", "dame la versión corta", "quiero controlar cuánto usas", or clear variants where the user is explicitly asking to control answer size or depth. DO NOT TRIGGER when: user has already specified a level in the current session (maintain it), the request is clearly a one-word answer, or "token" refers to auth/session/payment tokens rather than response size.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "token-budget-advisor" with this command: npx skills add affaan-m/everything-claude-code/affaan-m-everything-claude-code-token-budget-advisor

Token Budget Advisor (TBA)

Intercept the response flow to offer the user a choice about response depth before Claude answers.

When to Use

  • User wants to control how long or detailed a response is
  • User mentions tokens, budget, depth, or response length
  • User says "short version", "tldr", "brief", "al 25%", "exhaustive", etc.
  • Any time the user wants to choose depth/detail level upfront

Do not trigger when: user already set a level this session (maintain it silently), or the answer is trivially one line.

How It Works

Step 1 — Estimate input tokens

Use the repository's canonical context-budget heuristics to estimate the prompt's token count mentally.

Use the same calibration guidance as context-budget:

  • prose: words × 1.3
  • code-heavy or mixed/code blocks: chars / 4

For mixed content, use the dominant content type and keep the estimate heuristic.

Step 2 — Estimate response size by complexity

Classify the prompt, then apply the multiplier range to get the full response window:

ComplexityMultiplier rangeExample prompts
Simple3× – 8×"What is X?", yes/no, single fact
Medium8× – 20×"How does X work?"
Medium-High10× – 25×Code request with context
Complex15× – 40×Multi-part analysis, comparisons, architecture
Creative10× – 30×Stories, essays, narrative writing

Response window = input_tokens × mult_min to input_tokens × mult_max (but don’t exceed your model’s configured output-token limit).

Step 3 — Present depth options

Present this block before answering, using the actual estimated numbers:

Analyzing your prompt...

Input: ~[N] tokens  |  Type: [type]  |  Complexity: [level]  |  Language: [lang]

Choose your depth level:

[1] Essential   (25%)  ->  ~[tokens]   Direct answer only, no preamble
[2] Moderate    (50%)  ->  ~[tokens]   Answer + context + 1 example
[3] Detailed    (75%)  ->  ~[tokens]   Full answer with alternatives
[4] Exhaustive (100%)  ->  ~[tokens]   Everything, no limits

Which level? (1-4 or say "25% depth", "50% depth", "75% depth", "100% depth")

Precision: heuristic estimate ~85-90% accuracy (±15%).

Level token estimates (within the response window):

  • 25% → min + (max - min) × 0.25
  • 50% → min + (max - min) × 0.50
  • 75% → min + (max - min) × 0.75
  • 100% → max

Step 4 — Respond at the chosen level

LevelTarget lengthIncludeOmit
25% Essential2-4 sentences maxDirect answer, key conclusionContext, examples, nuance, alternatives
50% Moderate1-3 paragraphsAnswer + necessary context + 1 exampleDeep analysis, edge cases, references
75% DetailedStructured responseMultiple examples, pros/cons, alternativesExtreme edge cases, exhaustive references
100% ExhaustiveNo restrictionEverything — full analysis, all code, all perspectivesNothing

Shortcuts — skip the question

If the user already signals a level, respond at that level immediately without asking:

What they sayLevel
"1" / "25% depth" / "short version" / "brief answer" / "tldr"25%
"2" / "50% depth" / "moderate depth" / "balanced answer"50%
"3" / "75% depth" / "detailed answer" / "thorough answer"75%
"4" / "100% depth" / "exhaustive answer" / "full deep dive"100%

If the user set a level earlier in the session, maintain it silently for subsequent responses unless they change it.

Precision note

This skill uses heuristic estimation — no real tokenizer. Accuracy ~85-90%, variance ±15%. Always show the disclaimer.

Examples

Triggers

  • "Give me the short version first."
  • "How many tokens will your answer use?"
  • "Respond at 50% depth."
  • "I want the exhaustive answer, not the summary."
  • "Dame la version corta y luego la detallada."

Does Not Trigger

  • "What is a JWT token?"
  • "The checkout flow uses a payment token."
  • "Is this normal?"
  • "Complete the refactor."
  • Follow-up questions after the user already chose a depth for the session

Source

Standalone skill from TBA — Token Budget Advisor for Claude Code. Original project also ships a Python estimator script, but this repository keeps the skill self-contained and heuristic-only.

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Coding

coding-standards

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

backend-patterns

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

golang-patterns

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

frontend-patterns

No summary provided by upstream source.

Repository SourceNeeds Review