llm-speedtest

Ping major LLM providers in parallel and compare real API latency. Run with /ping

Safety Notice

This listing is from the official public ClawHub registry. Review SKILL.md and referenced scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "llm-speedtest" with this command: npx skills add chapati23/llm-speedtest

LLM Speedtest

Ping major LLM providers in parallel and compare real API latency (TTFT).

When to Use

  • User types /ping or asks about model latency/speed
  • Comparing provider response times
  • Checking if a specific provider is slow or down

How It Works

Runs scripts/ping.sh which:

  1. Retrieves API keys from pass shared/ (users may need to adapt key sourcing for their setup)
  2. Fires parallel curl requests to each provider with a minimal prompt ("hi", max_tokens=1)
  3. Measures total round-trip time per provider
  4. Sorts results by latency and displays with color badges

Output Format

Results are sorted fastest-to-slowest with color badges:

  • 🟢 < 2s — Fast
  • 🟡 2–5s — Normal
  • 🔴 5–30s — Slow
  • 30s — Timeout

Example:

⚡ Model Latency — 14:32

🟢 `Gemini       412ms`
🟢 `GPT-4o       623ms`
🟢 `Sonnet       891ms`
🟡 `Grok        2104ms`
🟡 `MiniMax     3210ms`
🟡 `Opus        4102ms`

_real API latency (TTFT)_

Models Tested

ProviderModel
AnthropicClaude Sonnet 4
AnthropicClaude Opus 4
OpenAIGPT-4o-mini
GoogleGemini 2.5 Flash
MiniMaxMiniMax-M1
xAIGrok 3 Mini Fast

Cost

~$0.0001 per run (1 token per model, cheapest tiers).

Note

This skill uses pass shared/ for API key retrieval. If you don't use pass, you'll need to adapt scripts/ping.sh to source keys from environment variables or another secret manager.

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.