OpenClaw Direct Setup
Double-click to install! No Docker required.
Quick Start
Windows: Double-click setup.bat or setup.ps1
Linux/Mac: Run bash setup.sh
The script will:
- Check if Node.js is installed
- Install OpenClaw globally
- Install ClawHub CLI and skills (pc-assistant, event-monitor)
- Create config with your LLM settings
- Start Ollama (if not running)
- Start the OpenClaw gateway
That's it!
Requirements
- Node.js - Download from https://nodejs.org (LTS version)
- npm - Comes with Node.js
What Gets Installed
| Component | Details |
|---|---|
| OpenClaw | npm install -g openclaw |
| ClawHub | npm install -g clawhub |
| Skills | pc-assistant, event-monitor |
| Config | ~/.openclaw/openclaw.json |
| Workspace | ~/.openclaw/workspace |
Security Features
- Localhost-only binding (127.0.0.1) - No external network access
- Admin/root detection - Blocks if running with elevated privileges
- Optional firewall rules - Blocks external access to gateway port
- Tailscale disabled by default - No account required
Configuration Options
During setup, you'll be asked for:
| Option | Default | Description |
|---|---|---|
| Port | 18789 | Gateway port |
| LLM Base URL | http://localhost:11434/v1 | Your LLM API endpoint |
| LLM Model | ServiceNow-AI/Apriel-1.6-15b-Thinker:Q4_K_M | Model ID |
Common LLM Setups
Ollama
LLM Base URL: http://localhost:11434/v1
LLM Model: llama3
LM Studio
LLM Base URL: http://localhost:1234/v1
LLM Model: llama-3-8b
vLLM
LLM Base URL: http://localhost:8000/v1
LLM Model: mistral-7b
OpenAI
LLM Base URL: https://api.openai.com/v1
LLM Model: gpt-4o
After installation, edit ~/.openclaw/openclaw.json to add your API key.
After Installation
- URL: http://localhost:18789
- Auth Token: Shown after installation
- Config:
~/.openclaw/openclaw.json
Commands
# Start gateway
openclaw gateway
# Stop gateway
openclaw gateway stop
# Check status
openclaw status
Uninstall
npm uninstall -g openclaw
rm -rf ~/.openclaw
Scripts
| File | Platform | Usage |
|---|---|---|
setup.bat | Windows | Double-click or run in CMD |
setup.ps1 | Windows | Right-click → Run with PowerShell |
setup.sh | Linux/Mac | bash setup.sh |
Problems?
- Node.js not found - Restart after installing Node.js
- Port in use - Choose a different port during setup
- LLM not connecting - Check your LLM is running and the URL is correct
- Ollama port conflict - Script detects if port 11434 is already in use
Version History
1.0.0
- Initial release
- Security hardening (localhost binding, admin detection)
- Auto-install pc-assistant and event-monitor skills
- Ollama port conflict detection
- Optional firewall rules