caching-strategies

Caching is the most commonly misapplied performance technique. The failure mode is not "cache too little" — it is "cache without an invalidation strategy and then discover the problem in production six months later when users complain about stale data that you cannot explain."

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "caching-strategies" with this command: npx skills add erichowens/some_claude_skills/erichowens-some-claude-skills-caching-strategies

Caching Strategies

Caching is the most commonly misapplied performance technique. The failure mode is not "cache too little" — it is "cache without an invalidation strategy and then discover the problem in production six months later when users complain about stale data that you cannot explain."

When to Use

✅ Use for:

  • Choosing which caching pattern fits a use case (cache-aside, write-through, write-behind)

  • Designing TTL values for different data freshness requirements

  • Implementing Redis caching patterns: sorted sets, pub/sub invalidation, Lua scripts

  • Configuring Cache-Control headers, ETags, and CDN behavior

  • Preventing cache stampedes via locking, probabilistic early expiry, or background refresh

  • Cache warming strategies for cold-start scenarios

  • Multi-tier cache design (in-memory L1, Redis L2, CDN L3)

❌ NOT for:

  • Database-internal query plan caching (handled by the database)

  • Python functools.lru_cache / JavaScript memoize utilities (pure function memoization)

  • CPU branch prediction or hardware cache tuning

  • Session storage (use dedicated session skill)

Which Caching Pattern?

flowchart TD Q1{Who writes to cache?} --> WA[Application writes] Q1 --> WC[Cache writes automatically] WA --> Q2{When does the cache get populated?} Q2 -->|On read miss| CA[Cache-Aside\n'Lazy loading'] Q2 -->|On every write| WT[Write-Through\n'Eager write'] WC --> Q3{Sync or async write-back?} Q3 -->|Sync — write completes when cache updates| WT Q3 -->|Async — write returns fast, flush later| WB[Write-Behind\n'Write-back'] CA --> N1{Is stale data OK\nfor a short period?} N1 -->|Yes| CA_USE[Use cache-aside\nwith TTL expiry] N1 -->|No| INVAL[Add explicit invalidation\nor use write-through] WT --> NOTE2[Good for read-heavy data\nthat changes infrequently] WB --> NOTE3[Good for write-heavy workloads\nRisk: data loss on crash]

Multi-Tier Cache Architecture

flowchart LR USER[User Request] --> CDN{CDN / Edge Cache\nL3 — 100ms+ saved} CDN -->|Cache hit| RESP[Response] CDN -->|Cache miss| LB[Load Balancer] LB --> APP[App Server] APP --> L1{In-Process Cache\nL1 — ~0ms} L1 -->|Hit| APP L1 -->|Miss| REDIS{Redis\nL2 — 1-5ms} REDIS -->|Hit| APP REDIS -->|Miss| DB[(Database\n10-100ms)] DB --> REDIS REDIS --> APP APP --> L1 APP --> CDN APP --> RESP

Tier Technology Latency Capacity Shared?

L1: In-process Node.js Map, Python dict, LRU-cache ~0ms Small (MB) No — per instance

L2: Distributed Redis, Memcached 1-5ms Large (GB) Yes — all instances

L3: Edge/CDN Cloudflare, Fastly, CloudFront 10-100ms Massive Yes — globally

Rule: Data mutates in one place first. Invalidation flows outward: DB → Redis → CDN. Never skip tiers in invalidation.

Cache-Aside Pattern (Most Common)

Application manages cache explicitly. On read: check cache, if miss fetch from DB, populate cache, return. On write: update DB, delete cache entry.

class UserCache { private redis: Redis; private readonly TTL_SECONDS = 300; // 5 minutes

async getUser(userId: string): Promise<User> { const key = user:${userId};

// 1. Check cache
const cached = await this.redis.get(key);
if (cached) return JSON.parse(cached);

// 2. Cache miss — fetch from source
const user = await db.users.findById(userId);
if (!user) throw new NotFoundError('User', userId);

// 3. Populate cache
await this.redis.setex(key, this.TTL_SECONDS, JSON.stringify(user));

return user;

}

async updateUser(userId: string, data: Partial<User>): Promise<User> { const user = await db.users.update(userId, data);

// 4. Invalidate — delete, don't update
// Updating in cache risks race conditions; let the next read repopulate
await this.redis.del(`user:${userId}`);

return user;

} }

When invalidation deletes vs overwrites: Delete is almost always correct. Overwriting in cache after a write creates a race: another request may have fetched the old value between your DB write and your cache write. Delete forces the next reader to fetch fresh.

Write-Through Pattern

Every write goes to cache and DB synchronously. Cache is always populated. Good for data that is written once and read many times.

async function createProduct(data: CreateProductInput): Promise<Product> { // Write to DB first (source of truth) const product = await db.products.create(data);

// Immediately populate cache — no future cache miss for this product const key = product:${product.id}; await redis.setex(key, 3600, JSON.stringify(product));

// Also invalidate list caches that include this product await redis.del('products:list:*'); // pattern delete via SCAN, see redis-patterns.md

return product; }

Trade-off: Higher write latency (two writes per operation). Wasted cache space for items that are never read again after creation. Best for data with high read:write ratio.

TTL Design

TTL is not a cache invalidation strategy — it is a staleness budget. Design TTLs based on data volatility and acceptable staleness:

Data Type TTL Rationale

User session token Match session expiry Security requirement

User profile (name, avatar) 5-15 minutes Changes rarely; short enough for responsiveness

Product catalog 1-4 hours Changes occasionally; acceptable lag

Inventory counts 30 seconds Changes frequently; short but not zero

Exchange rates 60 seconds Regulatory; must not be too stale

Static config / feature flags 60 seconds + pub/sub invalidation Needs push invalidation on change

Computed aggregates (daily stats) Until next computation Explicit invalidation on recalculate

TTL jitter: When many keys have the same TTL, they expire simultaneously, causing a thundering herd. Add random jitter:

const jitter = Math.floor(Math.random() * 60); // 0-60 seconds await redis.setex(key, baseTtl + jitter, value);

Cache Stampede Prevention

A stampede (also: dog-pile, thundering herd) occurs when many requests simultaneously miss an expired cache key and all rush to compute or fetch the value.

Strategy 1: Probabilistic Early Expiry (XFetch)

Re-fetch before expiry with probability proportional to how close the key is to expiring:

async function getWithEarlyExpiry<T>( key: string, fetcher: () => Promise<T>, ttlSeconds: number, beta = 1.0 ): Promise<T> { const entry = await redis.get(key + ':meta'); if (entry) { const { value, expiresAt, fetchDurationMs } = JSON.parse(entry); const now = Date.now(); const ttlRemaining = expiresAt - now; // Fetch early if within probabilistic window const shouldRefetch = ttlRemaining < beta * fetchDurationMs * Math.log(Math.random()); if (!shouldRefetch) return value; }

// Fetch and cache const start = Date.now(); const value = await fetcher(); const fetchDurationMs = Date.now() - start; const expiresAt = Date.now() + ttlSeconds * 1000;

await redis.setex(key + ':meta', ttlSeconds, JSON.stringify({ value, expiresAt, fetchDurationMs })); return value; }

Strategy 2: Mutex Lock on Miss

Only one worker recomputes the value; others wait on the lock or return stale data:

async function getWithLock<T>( key: string, fetcher: () => Promise<T>, ttl: number ): Promise<T> { const cached = await redis.get(key); if (cached) return JSON.parse(cached);

const lockKey = lock:${key}; const lockAcquired = await redis.set(lockKey, '1', 'NX', 'PX', 5000); // 5s TTL

if (!lockAcquired) { // Another worker is computing — poll briefly then return stale or throw await sleep(100); const retried = await redis.get(key); if (retried) return JSON.parse(retried); throw new Error('Cache unavailable'); }

try { const value = await fetcher(); await redis.setex(key, ttl, JSON.stringify(value)); return value; } finally { await redis.del(lockKey); } }

Consult references/redis-patterns.md for the Lua-atomic version of this lock (prevents lock release by wrong client).

Anti-Patterns

Anti-Pattern: Cache Everything Forever

Novice: "Caching makes things fast. Set TTL to 0 (no expiry) or a year to maximize cache hit rate."

Expert: Unbounded caches are memory leaks with extra steps. They also guarantee stale data — users see prices, permissions, and content from months ago. Production incidents traced to "why is this user seeing the old plan limit" are almost always cache-forever bugs.

// Wrong — no expiry means the cache grows forever await redis.set(user:${id}, JSON.stringify(user)); // no TTL

// Right — every cache entry has a maximum lifetime await redis.setex(user:${id}, 300, JSON.stringify(user)); // 5 minutes

Python equivalent:

Wrong

redis.set(f"user:{id}", json.dumps(user))

Right

redis.setex(f"user:{id}", 300, json.dumps(user))

Detection: redis.set(key, value) without EX /PX /EXAT options. Redis TTL key returning -1 for cache keys. Memory growth over time with no plateau.

Timeline: This has always been wrong, but the Redis default of no-expiry makes it easy to do accidentally. Redis 7.0 (2022) introduced key eviction policies as default, reducing severity — but you still get stale data.

Anti-Pattern: No Invalidation Strategy

Novice: "I'll set a short TTL and the stale data problem solves itself."

Expert: TTL-only invalidation means every change to data has a propagation delay equal to the TTL. For some data (user roles, permissions, prices after a sale ends) that lag is unacceptable. Worse: this creates an implicit contract that is never documented, and teams later increase the TTL for performance without realizing they just made the staleness window much larger.

// Problem: user loses admin role, but can still access admin routes for 5 minutes await redis.setex(user:permissions:${id}, 300, JSON.stringify(permissions));

// Right: invalidate explicitly on change async function revokeAdminRole(userId: string) { await db.userRoles.delete(userId, 'admin'); await redis.del(user:permissions:${userId}); // immediate invalidation // Also publish to notify other app instances to clear L1 caches await redis.publish('permissions:invalidated', userId); }

LLM mistake: LLMs frequently omit invalidation logic in code generation because it is invisible in simple cache-aside examples. Every tutorial shows "set on write," few show "delete on update."

Detection: Cache sets with no corresponding deletes in write paths. TTL as the only eviction mechanism for user-controlled data (roles, permissions, settings). No DEL , UNLINK , or pub/sub events in the codebase's update handlers.

References

  • references/redis-patterns.md — Consult for Redis-specific patterns: sorted sets for leaderboards, Lua atomic operations, pub/sub cache invalidation, SCAN-based key deletion, pipeline batching

  • references/http-caching.md — Consult for browser caching: Cache-Control directives, ETags, Vary headers, CDN configuration, service worker caching strategies

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

General

video-processing-editing

No summary provided by upstream source.

Repository SourceNeeds Review
General

cv-creator

No summary provided by upstream source.

Repository SourceNeeds Review
General

mobile-ux-optimizer

No summary provided by upstream source.

Repository SourceNeeds Review
General

personal-finance-coach

No summary provided by upstream source.

Repository SourceNeeds Review