mirror of
https://github.com/openclaw/openclaw.git
synced 2026-03-20 14:30:57 +00:00
Redirects: - /cron now goes directly to /automation/cron-jobs (was chaining via /cron-jobs) - /model and /model/ now go directly to /concepts/models (was chaining via /models) Duplicate titles disambiguated (6 of 7 - Logging is orphaned): - Health Checks (macOS), Skills (macOS), Voice Wake (macOS), WebChat (macOS) - General Troubleshooting (help/ vs gateway/) - Provider Directory (providers/index vs concepts/model-providers) Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
2.0 KiB
2.0 KiB
summary, read_when, title
| summary | read_when | title | ||
|---|---|---|---|---|
| Model providers (LLMs) supported by OpenClaw |
|
Provider Directory |
Model Providers
OpenClaw can use many LLM providers. Pick a provider, authenticate, then set the
default model as provider/model.
Looking for chat channel docs (WhatsApp/Telegram/Discord/Slack/Mattermost (plugin)/etc.)? See Channels.
Quick start
- Authenticate with the provider (usually via
openclaw onboard). - Set the default model:
{
agents: { defaults: { model: { primary: "anthropic/claude-opus-4-6" } } },
}
Provider docs
- Amazon Bedrock
- Anthropic (API + Claude Code CLI)
- Cloudflare AI Gateway
- GLM models
- Hugging Face (Inference)
- Kilocode
- LiteLLM (unified gateway)
- MiniMax
- Mistral
- Moonshot AI (Kimi + Kimi Coding)
- NVIDIA
- Ollama (cloud + local models)
- OpenAI (API + Codex)
- OpenCode (Zen + Go)
- OpenRouter
- Qianfan
- Qwen (OAuth)
- Together AI
- Vercel AI Gateway
- Venice (Venice AI, privacy-focused)
- vLLM (local models)
- xAI
- Xiaomi
- Z.AI
Transcription providers
Community tools
- Claude Max API Proxy - Community proxy for Claude subscription credentials (verify Anthropic policy/terms before use)
For the full provider catalog (xAI, Groq, Mistral, etc.) and advanced configuration, see Model providers.