tanstack-pacer

TanStack Pacer provides a unified, type-safe toolkit for controlling function execution timing. It offers class-based APIs, factory functions, and React hooks for debouncing, throttling, rate limiting, queuing, and batching.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "tanstack-pacer" with this command: npx skills add tanstack-skills/tanstack-skills/tanstack-skills-tanstack-skills-tanstack-pacer

Overview

TanStack Pacer provides a unified, type-safe toolkit for controlling function execution timing. It offers class-based APIs, factory functions, and React hooks for debouncing, throttling, rate limiting, queuing, and batching.

Core: @tanstack/pacer

React: @tanstack/react-pacer

Status: Beta

Installation

npm install @tanstack/pacer npm install @tanstack/react-pacer # React hooks

Debouncing

Delays execution until after a period of inactivity.

Class API

import { Debouncer } from '@tanstack/pacer'

const debouncer = new Debouncer( (query: string) => fetchSearchResults(query), { wait: 300, // ms of inactivity before execution leading: false, // Execute on leading edge (default: false) trailing: true, // Execute on trailing edge (default: true) maxWait: 1000, // Force execution after 1s of continuous calls enabled: true, onExecute: (result) => console.log(result), } )

debouncer.maybeExecute('search term') debouncer.cancel() debouncer.getExecutionCount() debouncer.setOptions({ wait: 500 }) // Dynamic reconfiguration

Factory Function

import { debounce } from '@tanstack/pacer'

const debouncedSearch = debounce( (query: string) => fetchResults(query), { wait: 300 } )

debouncedSearch('term') debouncedSearch.cancel()

React Hooks

import { useDebouncer, useDebouncedCallback, useDebouncedState, useDebouncedValue, } from '@tanstack/react-pacer'

// Full debouncer instance const debouncer = useDebouncer(fn, { wait: 300 })

// Simple debounced function const debouncedFn = useDebouncedCallback(fn, { wait: 300 })

// Debounced state management const [debouncedValue, setValue] = useDebouncedState(initialValue, { wait: 300 })

// Debounced reactive value const debouncedValue = useDebouncedValue(reactiveValue, { wait: 300 })

Throttling

Limits execution to at most once per interval.

Class API

import { Throttler } from '@tanstack/pacer'

const throttler = new Throttler( (position: { x: number; y: number }) => updatePosition(position), { wait: 100, // Minimum interval between executions leading: true, // Execute immediately on first call (default: true) trailing: true, // Execute after interval with last args (default: true) enabled: true, onExecute: (result) => console.log(result), } )

throttler.maybeExecute({ x: 100, y: 200 }) throttler.cancel()

React Hooks

import { useThrottler, useThrottledCallback, useThrottledState, useThrottledValue, } from '@tanstack/react-pacer'

const throttledFn = useThrottledCallback(handleScroll, { wait: 100 }) const [throttledPos, setPos] = useThrottledState({ x: 0, y: 0 }, { wait: 100 })

Rate Limiting

Controls execution with a maximum count within a time window.

Class API

import { RateLimiter } from '@tanstack/pacer'

const limiter = new RateLimiter( async (endpoint: string) => fetch(endpoint).then(r => r.json()), { limit: 10, // Max executions per window window: 60000, // Time window in ms (60s) enabled: true, onExecute: (result) => console.log(result), onReject: (...args) => console.warn('Rate limited:', args), } )

limiter.maybeExecute('/api/data') // Rejected if limit exceeded limiter.getExecutionCount() limiter.getRejectionCount()

React Hooks

import { useRateLimiter, useRateLimitedCallback, useRateLimitedState, useRateLimitedValue, } from '@tanstack/react-pacer'

const rateLimitedFn = useRateLimitedCallback(apiCall, { limit: 5, window: 1000 })

Queuing

Sequential execution with configurable concurrency.

import { Queue } from '@tanstack/pacer'

const queue = new Queue({ concurrency: 1, // Max concurrent tasks started: true, // Start processing immediately })

queue.add(() => uploadFile(file1)) queue.add(() => uploadFile(file2))

queue.start() queue.pause() queue.clear() queue.getSize() // Pending count queue.getPending() // Currently executing count

Batching

Groups calls for combined processing.

import { Batcher } from '@tanstack/pacer'

const batcher = new Batcher( (items: LogEntry[]) => sendBatchToServer(items), { maxSize: 50, // Auto-flush at 50 items wait: 1000, // Auto-flush after 1s } )

batcher.add(logEntry) // Accumulates batcher.flush() // Manual flush batcher.getSize() // Current batch size batcher.clear() // Discard batch

Async Variants

import { AsyncDebouncer, asyncDebounce, AsyncThrottler, asyncThrottle } from '@tanstack/pacer'

const asyncDebouncer = new AsyncDebouncer( async (query: string) => { const response = await fetch(/api/search?q=${query}) return response.json() }, { wait: 300 } )

// React async hooks import { useAsyncDebouncer, useAsyncThrottler } from '@tanstack/react-pacer'

Choosing the Right Utility

Scenario Utility Why

Search input Debouncer Wait for user to stop typing

Scroll events Throttler Periodic updates during activity

API protection RateLimiter Hard limit on call frequency

File uploads Queue Sequential processing

Analytics events Batcher Group for efficiency

Network requests AsyncDebouncer Handle abort/retry

Leading vs Trailing Edge

  • Leading (leading: true ): Execute immediately, suppress until wait expires. Good for button clicks.

  • Trailing (trailing: true ): Execute after activity stops. Good for search inputs.

  • Both: Execute immediately AND after final wait. Good for scroll throttling.

Best Practices

  • Use maxWait with debouncing to guarantee execution during continuous activity

  • Use async variants for network requests (handle abort/cancellation)

  • React hooks handle cleanup automatically - no manual teardown needed

  • Use setOptions for dynamic reconfiguration (e.g., reducing wait for power users)

  • Compose utilities for complex scenarios (rate-limited queue)

  • Use onReject on RateLimiter to inform users when they're rate limited

Common Pitfalls

  • Using debounce when you need throttle (debounce waits for inactivity, throttle guarantees periodic execution)

  • Not using maxWait with debounce for long-running continuous events

  • Creating new instances on every render (use hooks or module-level)

  • Forgetting cleanup in non-React environments (call cancel() )

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

General

tanstack-query

No summary provided by upstream source.

Repository SourceNeeds Review
General

tanstack-table

No summary provided by upstream source.

Repository SourceNeeds Review
General

tanstack-router

No summary provided by upstream source.

Repository SourceNeeds Review