mirror of
https://github.com/openclaw/openclaw.git
synced 2026-03-26 01:11:37 +00:00
- Sort providers alphabetically in docs.json nav - Sort channels alphabetically in docs.json nav (slack before synology-chat) - Add install/migrating-matrix to Maintenance nav section (was orphaned) - Remove zh-CN/plugins/architecture from nav (file does not exist) - Add Voice Call to channels index page - Add missing providers to providers index (DeepSeek, GitHub Copilot, OpenCode Go, Synthetic) - Sort providers index alphabetically - Update stale claude-3-5-sonnet model reference to claude-sonnet-4-6 in webhook docs
2.4 KiB
2.4 KiB
summary, read_when, title
| summary | read_when | title | ||
|---|---|---|---|---|
| Model providers (LLMs) supported by OpenClaw |
|
Provider Directory |
Model Providers
OpenClaw can use many LLM providers. Pick a provider, authenticate, then set the
default model as provider/model.
Looking for chat channel docs (WhatsApp/Telegram/Discord/Slack/Mattermost (plugin)/etc.)? See Channels.
Quick start
- Authenticate with the provider (usually via
openclaw onboard). - Set the default model:
{
agents: { defaults: { model: { primary: "anthropic/claude-opus-4-6" } } },
}
Provider docs
- Amazon Bedrock
- Anthropic (API + Claude Code CLI)
- Cloudflare AI Gateway
- DeepSeek
- GitHub Copilot
- GLM models
- Google (Gemini)
- Groq (LPU inference)
- Hugging Face (Inference)
- Kilocode
- LiteLLM (unified gateway)
- MiniMax
- Mistral
- Model Studio (Alibaba Cloud)
- Moonshot AI (Kimi + Kimi Coding)
- NVIDIA
- Ollama (cloud + local models)
- OpenAI (API + Codex)
- OpenCode
- OpenCode Go
- OpenRouter
- Perplexity (web search)
- Qianfan
- Qwen (OAuth)
- SGLang (local models)
- Synthetic
- Together AI
- Venice (Venice AI, privacy-focused)
- Vercel AI Gateway
- vLLM (local models)
- Volcengine (Doubao)
- xAI
- Xiaomi
- Z.AI
Transcription providers
Community tools
- Claude Max API Proxy - Community proxy for Claude subscription credentials (verify Anthropic policy/terms before use)
For the full provider catalog (xAI, Groq, Mistral, etc.) and advanced configuration, see Model providers.