LLM Deploy
在 GPU 服务器上部署 LLM 模型服务(vLLM)。支持多服务器配置,自动检查 GPU 和端口占用,一键部署流行的开源大语言模型。
Save 30% GPU cost with architecture-aware AI advisor. Powered by the world's first RTX 5090 Energy Paradox study. 93+ empirical measurements, real-time dolla...
This listing is from the official public ClawHub registry. Review SKILL.md and referenced scripts before running.
Install skill "EcoCompute — LLM Energy Efficiency Advisor" with this command: npx skills add ecocompute
This source entry does not include full markdown content beyond metadata.
This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.
Related by shared tags or category signals.
在 GPU 服务器上部署 LLM 模型服务(vLLM)。支持多服务器配置,自动检查 GPU 和端口占用,一键部署流行的开源大语言模型。
Comprehensive system for designing, testing, optimizing, and managing clear, role-aware, actionable, focused, and testable prompts for AI models.
Find carbon emission factors for any material or process. 1M+ LCA datasets (HiQLCD, Ecoinvent, CarbonMinds). AI-powered BOM carbon footprint calculation and...