lm-studio-discord

Connect a local LM Studio model directly to Discord as a lightweight chat bot. Use when you want to expose a local LLM (running via LM Studio on CPU) to Discord without OpenClaw's overhead. Bypasses the main OpenClaw gateway for speed on CPU-only machines.

Safety Notice

This listing is from the official public ClawHub registry. Review SKILL.md and referenced scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "lm-studio-discord" with this command: npx skills add dq-stack/lm-studio-discord

LM Studio Discord Bot

A minimal Discord bot that routes messages directly to LM Studio — no OpenClaw gateway in the path.

When to Use

  • LM Studio model is fast in direct chat but slow through OpenClaw
  • CPU-only setup where OpenClaw's tool system adds too much latency
  • You want a simple Discord bot without the full OpenClaw feature set

How It Works

Discord → bot.js → LM Studio API → bot.js → Discord

The bot uses Discord.js to receive messages and axios to call LM Studio's /v1/chat/completions endpoint directly. No tools, no workspace files, no system prompt overhead.

Setup

1. Create the bot project

mkdir lm-studio-discord-bot
cd lm-studio-discord-bot
npm init -y
npm install discord.js axios

2. Create bot.js

See scripts/bot-template.js for the ready-to-use template.

3. Configure

Edit these constants in bot.js:

const DISCORD_TOKEN = 'YOUR_DISCORD_BOT_TOKEN';
const LM_STUDIO_URL = 'http://127.0.0.1:1234/v1/chat/completions';
const MODEL = 'qwen2-0.5b-instruct';  // Must match loaded model in LM Studio
const GUILD_ID = 'YOUR_DISCORD_SERVER_ID';

To get your bot token: https://discord.com/developers/applications

To get the Guild ID: Enable Developer Mode in Discord → Right-click your server → Copy ID

4. Add bot to your server

https://discord.com/oauth2/authorize?client_id=BOT_CLIENT_ID&permissions=1024&scope=bot

Replace BOT_CLIENT_ID with your bot's Application ID from the Discord Developer Portal.

5. Run

node bot.js

The bot will log in and respond to messages in any channel of the configured guild.

Configuration Tips

ParameterDefaultNotes
max_tokens512Lower = faster, less verbose
timeout60000msIncrease if slow on CPU
streamfalseSet true for streaming replies

Known Limitations

  • No tool access (file read/write, web search, etc.)
  • No conversation memory — each message is stateless
  • No slash commands
  • Single model per bot instance

Extending

To add conversation history, modify the messages array in the axios call to include prior exchanges. Note: this increases token usage and context window pressure.

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Automation

飞书任务Bot身份

飞书任务(Bot身份):使用Bot/应用身份管理飞书任务,创建任务、查询任务列表、更新状态、分配成员等。本Skill专门使用v1 API,Bot身份可直接调用,解决了lark-task官方Skill使用v2接口无法支持Bot身份的问题。当需要以Bot身份(应用身份)操作任务时使用本Skill。

Registry SourceRecently Updated
1080Profile unavailable
Automation

Cross-Platform Memory Bridge

Injects recent conversations from Telegram and Discord into the OpenClaw gateway session context. Enables the agent to remember and reference cross-platform...

Registry Source
1030Profile unavailable
Automation

HITL Protocol — Workflows and Human Decisions for Autonomous Agents

HITL Protocol — the open standard for human decisions in autonomous agent workflows. When a website or API needs human input, it returns HTTP 202 with a revi...

Registry Source
5562Profile unavailable
Automation

OpenClaw Profanity Plugin

Content moderation plugin for OpenClaw/Moltbot AI agents. Use when building chatbots that need profanity filtering, moderating user messages in Discord/Slack/Telegram bots, or adding content moderation to OpenClaw agents.

Registry Source
1.6K1Profile unavailable