local-inference-context

Context management for self-hosted LLM backends (llama.cpp, Ollama). Prevents mid-task 503 errors and context overflows caused by VRAM-limited KV caches. Use instead of generic context skills when running a local inference server. Complements context-recovery for post-compaction scenarios.

Safety Notice

This listing is from the official public ClawHub registry. Review SKILL.md and referenced scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "local-inference-context" with this command: npx skills add joekravelli/local-inference-context

Local Inference Context

Generic context skills assume a reliable, large-context cloud provider. Local backends (llama.cpp, Ollama) have a different failure profile: the KV cache is bounded by VRAM, the server can return 503 before OpenClaw's compaction logic triggers, and the compaction model is the same overloaded local model. This skill addresses that reality.


Why local backends fail differently

Cloud providerLocal llama.cpp / Ollama
Context limit is a soft API error, OpenClaw retries after compactionKV-cache fills up, server returns 503 or context length exceeded mid-request
Compaction uses same model, which is always availableCompaction uses same overloaded local model — may also fail
Context window is exactly what the API reportsEffective context = min(configured --ctx-size, available VRAM for KV cache)
No idle slot evictionIdle slots can be evicted; server returns "Loading model" 503 on next request

The practical consequence: on a GPU-constrained setup (e.g. a 24 GB card running a 27B Q5 model), the usable KV-cache budget is roughly 5–8 GB. At 32k tokens configured context, that fills up faster than the configured limit suggests. Treat 50 % fill as amber and 70 % as red — not 60/80 %.


Calibrating your effective context budget

Before a long session, run this once to understand your actual headroom:

# Check VRAM headroom
nvidia-smi --query-gpu=memory.used,memory.free,memory.total \
  --format=csv,noheader,nounits

# Check llama.cpp slot state
curl -s http://localhost:8081/slots | python3 -m json.tool

If memory.free is less than 4 GB, treat the session as already amber regardless of what /status reports. Log the result to memory:

VRAM free: X MB — effective context budget: reduced

Thresholds for local backends

Fill levelStateAction
< 50 %GreenProceed normally
50–69 %AmberTrim tool outputs, flush key facts to memory
70–84 %RedCheckpoint, offer /compact before continuing
≥ 85 %CriticalStop expanding. Compact or /new before next tool call

Check /status at session start and after any tool call that returns more than ~200 lines of output.


Recognising a local backend failure

These are server-side errors, not OpenClaw compaction events. They require a different response than a normal context overflow:

SignalMeaning
HTTP 503 with body "loading model"Idle slot was evicted; model is reloading. Wait 10–30 s, then retry once.
HTTP 503 with body "no slot available"All slots busy or KV cache full. Do NOT retry immediately — compact first.
context length exceeded in errorHard KV-cache overflow. Compact or start /new before any retry.
Sudden very slow response then timeoutKV cache thrashing — reduce context before next request.

Never retry a 503 "no slot available" or context overflow without first reducing context. Retrying makes the problem worse by sending the same oversized payload again.


Pre-task checklist for long operations

Before any task you expect to span more than 4 turns (file edits, debugging sessions, multi-step setups):

  1. Run /status — note current fill %.
  2. Check nvidia-smi if fill is already above 40 %.
  3. Estimate token cost of the task:
    • Each file read ≈ 500–3000 tokens depending on file size
    • Each exec result ≈ 200–1500 tokens
    • Each web_fetch ≈ 1000–4000 tokens
  4. If estimated total would push past 70 %, split into phases and tell the user upfront.

Amber state (50–69 %): lean tool hygiene

Apply these habits to every tool call in amber state:

# Instead of reading entire files:
sed -n '1,50p' /path/to/file          # first 50 lines
grep -n "error\|warn\|fail" logfile   # targeted grep
tail -100 /var/log/syslog             # recent entries only

# Instead of verbose exec output:
some-command 2>&1 | tail -30
systemctl status service --no-pager --lines=20

# Summarise large outputs in one sentence, then discard them:
# "Command succeeded. Key values: port=8081, pid=12345"

Write key values to memory immediately after each tool call — do not rely on them surviving a compaction summary intact.


Red state (70–84 %): checkpoint before continuing

  1. Write a checkpoint to memory now:
## Checkpoint [timestamp]
Status: [what is done]
Pending: [what is next]
Critical values: [file paths, ports, error codes, config keys]
  1. Tell the user:

⚠️ Context at ~N % (local backend — conservative threshold). I've saved progress to memory. Recommend /compact Focus on [task] before continuing. Or /new for a clean session.

  1. If continuing, use /compact Focus on <current task> — not bare /compact. The local model needs a focused instruction to produce a useful summary under memory pressure.

Critical state (≥ 85 %): stop and recover

Do not issue any more tool calls that expand context.

  1. Write the checkpoint (see above).
  2. Send the user a recovery message:
🛑 Context critical (~N %). Stopping to prevent a server error.

Done: [X]
Pending: [Y]
Key info: [Z]

Options:
  /compact Focus on [task]   — summarise and continue
  /new                       — fresh session (I'll reload from memory)
  1. Wait for the user to choose. Do not attempt to continue on your own.

After a 503 or context-overflow error

If the server already returned an error before you could act:

  1. Do not panic and do not retry the same request.
  2. Check the error type:
    • "loading model" → wait 15–30 s, then retry once with a minimal message.
    • "no slot available" or context length exceeded → compact first.
  3. Run /compact Focus on [what you were doing].
  4. After compaction, verify the slot is ready:
    curl -s http://localhost:8081/health
    # expect: {"status":"ok"}
    
  5. Re-read any file paths or config values from memory or disk — do not trust the compaction summary to have preserved them verbatim.
  6. Resume with a short, targeted first message to re-establish the session before loading more context.

Compaction model — required, not optional

Without a dedicated compaction model, OpenClaw uses the same local model for summarisation — the identical model whose KV cache just caused the overflow. This means compaction will likely fail or produce a degraded summary. A separate compaction model is a prerequisite for this skill to work reliably, not an optional optimisation.

The compaction model should run on a different machine or a second inference instance with its own memory budget. It does not need to be powerful — it only needs to summarise text faithfully and follow instructions. A 7B–8B model is sufficient.

Recommended model: qwen2.5:7b via Ollama (fits in ~5 GB RAM/VRAM, fast, excellent at summarisation and instruction-following). Fallback if speed is critical: llama3.2:3b (~2 GB).

{
  "agents": {
    "defaults": {
      "compaction": {
        "model": "ollama/qwen2.5:7b",
        "notifyUser": true,
        "memoryFlush": {
          "model": "ollama/qwen2.5:7b"
        }
      }
    }
  },
  "providers": {
    "ollama": {
      "baseUrl": "http://<COMPACTION-SERVER-IP>:11434"
    }
  }
}

Without this configuration, the skill provides partial benefit only: the conservative thresholds and lean tool habits reduce overflow frequency, but cannot recover reliably once an overflow occurs.


Slash command reference

CommandWhen to use
/statusCheck fill % — use at session start and after large tool outputs
/context listSee which injected files and skills consume the most tokens
/compact Focus on <topic>Guided compaction — always specify focus on a local backend
/newClean slate — fastest recovery when context is critical
/usage tokensPer-reply token counter — useful for calibrating estimates

Relationship to other skills

SkillWhen to use instead
context-recoveryAfter compaction on any backend — recovers lost context via channel history
context-budgetingCloud providers or stable local setups — heartbeat-based GC at >80 %
context-clean-upDiagnosing chronic context bloat — ranked offender audit
context-anchorPost-compaction orientation via memory file scan

Use local-inference-context before problems occur and context-recovery after compaction if context was lost.

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

General

Eyes

全球热点事件监控与影响分析。覆盖战争冲突、地缘摩擦、重大政策、疫情、自然灾害、创新技术等可能影响经济、市场和投资的事件,并按行业、汇率、大宗商品链路分析影响。也用于 Cron 定时推送热点摘要(早8点开盘前瞻/晚8点收盘复盘/整点扫描)。

Registry SourceRecently Updated
General

Send Md As

在即时通讯 app 中以优雅图片形式展示 Markdown。支持标题、代码高亮(行号、Monokai)、LaTeX 公式、Mermaid 图表、表格、列表。4 种色彩主题,智能分页。渲染阶段零 CDN 依赖(安装时需网络下载依赖)。| Render Markdown as a polished image for...

Registry SourceRecently Updated
General

Rails CI Fixer

Fix failing CI on Rails PRs using a tiered escalation loop. Use this skill whenever a Rails pull request has failing CI — RSpec failures, RuboCop offenses, m...

Registry SourceRecently Updated
4540djc00p
General

Interview Story Coach

Helps structure authentic behavioral interview stories using STAR/CAR with clear, concise answers, follow-up questions, and rehearsal tips.

Registry SourceRecently Updated