docs: refresh qwen media and config refs

This commit is contained in:
Peter Steinberger
2026-04-04 19:42:13 +01:00
parent b1279b0db3
commit 879d45a56c
7 changed files with 25 additions and 14 deletions

View File

@@ -1478,7 +1478,7 @@ at 12:00 PM PT / 8:00 PM BST**. Anthropic says Claude subscription limits no
longer cover OpenClaw, and Claude CLI usage in OpenClaw now requires **Extra
Usage** billed separately from the subscription. For production, prefer an
Anthropic API key or another supported subscription-style provider such as
OpenAI Codex, Alibaba Cloud Model Studio Coding Plan, MiniMax Coding Plan, or
OpenAI Codex, Qwen Cloud Coding Plan, MiniMax Coding Plan, or
Z.AI / GLM Coding Plan.
Anthropic Claude CLI migration:

View File

@@ -30,6 +30,7 @@ Related:
falls back to `agents.defaults.imageModel`, then the resolved session/default
model.
- `agents.defaults.imageGenerationModel` is used by the shared image-generation capability. If omitted, `image_generate` can still infer an auth-backed provider default. It tries the current default provider first, then the remaining registered image-generation providers in provider-id order. If you set a specific provider/model, also configure that provider's auth/API key.
- `agents.defaults.videoGenerationModel` is used by the shared video-generation capability. If omitted, video-generation providers can still use their own default model selection; if you set a specific provider/model, configure that provider's auth/API key too.
- Per-agent defaults can override `agents.defaults.model` via `agents.list[].model` plus bindings (see [/concepts/multi-agent](/concepts/multi-agent)).
## Quick model policy
@@ -55,6 +56,7 @@ subscription** (OAuth) and **Anthropic** (API key or Claude CLI).
- `agents.defaults.imageModel.primary` and `agents.defaults.imageModel.fallbacks`
- `agents.defaults.pdfModel.primary` and `agents.defaults.pdfModel.fallbacks`
- `agents.defaults.imageGenerationModel.primary` and `agents.defaults.imageGenerationModel.fallbacks`
- `agents.defaults.videoGenerationModel.primary` and `agents.defaults.videoGenerationModel.fallbacks`
- `agents.defaults.models` (allowlist + aliases + provider params)
- `models.providers` (custom providers written into `models.json`)

View File

@@ -952,6 +952,10 @@ Time format in system prompt. Default: `auto` (OS preference).
primary: "openai/gpt-image-1",
fallbacks: ["google/gemini-3.1-flash-image-preview"],
},
videoGenerationModel: {
primary: "qwen/wan2.6-t2v",
fallbacks: ["qwen/wan2.6-i2v"],
},
pdfModel: {
primary: "anthropic/claude-opus-4-6",
fallbacks: ["openai/gpt-5.4-mini"],
@@ -982,6 +986,10 @@ Time format in system prompt. Default: `auto` (OS preference).
- Typical values: `google/gemini-3.1-flash-image-preview` for native Gemini image generation, `fal/fal-ai/flux/dev` for fal, or `openai/gpt-image-1` for OpenAI Images.
- If you select a provider/model directly, configure the matching provider auth/API key too (for example `GEMINI_API_KEY` or `GOOGLE_API_KEY` for `google/*`, `OPENAI_API_KEY` for `openai/*`, `FAL_KEY` for `fal/*`).
- If omitted, `image_generate` can still infer an auth-backed provider default. It tries the current default provider first, then the remaining registered image-generation providers in provider-id order.
- `videoGenerationModel`: accepts either a string (`"provider/model"`) or an object (`{ primary, fallbacks }`).
- Used by the shared video-generation capability.
- Typical values: `qwen/wan2.6-t2v` or another registered video-generation provider/model.
- If you select a provider/model directly, configure the matching provider auth/API key too.
- `pdfModel`: accepts either a string (`"provider/model"`) or an object (`{ primary, fallbacks }`).
- Used by the `pdf` tool for model routing.
- If omitted, the PDF tool falls back to `imageModel`, then to the resolved session/default model.

View File

@@ -587,7 +587,7 @@ Quick answers plus deeper troubleshooting for real-world setups (local dev, VPS,
explicitly supported for external tools like OpenClaw.
OpenClaw also supports other hosted subscription-style options including
**Alibaba Cloud Model Studio Coding Plan**, **MiniMax Coding Plan**, and
**Qwen Cloud Coding Plan**, **MiniMax Coding Plan**, and
**Z.AI / GLM Coding Plan**.
Docs: [Anthropic](/providers/anthropic), [OpenAI](/providers/openai),

View File

@@ -160,7 +160,7 @@ working option**:
- Bundled fallback order:
- Audio: OpenAI → Groq → Deepgram → Google → Mistral
- Image: OpenAI → Anthropic → Google → MiniMax → MiniMax Portal → Z.AI
- Video: Google → Moonshot
- Video: Google → Qwen → Moonshot
To disable auto-detection, set:
@@ -202,6 +202,7 @@ lists, OpenClaw can infer defaults:
- `moonshot`: **image + video**
- `openrouter`: **image**
- `google` (Gemini API): **image + audio + video**
- `qwen`: **image + video**
- `mistral`: **audio**
- `zai`: **image**
- `groq`: **audio**
@@ -214,11 +215,11 @@ If you omit `capabilities`, the entry is eligible for the list it appears in.
## Provider support matrix (OpenClaw integrations)
| Capability | Provider integration | Notes |
| ---------- | -------------------------------------------------------------------------------- | ---------------------------------------------------------------------------------------------------------------------------------------- |
| Image | OpenAI, OpenRouter, Anthropic, Google, MiniMax, Moonshot, Z.AI, config providers | Vendor plugins register image support; MiniMax and MiniMax OAuth both use `MiniMax-VL-01`; image-capable config providers auto-register. |
| Audio | OpenAI, Groq, Deepgram, Google, Mistral | Provider transcription (Whisper/Deepgram/Gemini/Voxtral). |
| Video | Google, Moonshot | Provider video understanding via vendor plugins. |
| Capability | Provider integration | Notes |
| ---------- | -------------------------------------------------------------------------------------- | ---------------------------------------------------------------------------------------------------------------------------------------- |
| Image | OpenAI, OpenRouter, Anthropic, Google, MiniMax, Moonshot, Qwen, Z.AI, config providers | Vendor plugins register image support; MiniMax and MiniMax OAuth both use `MiniMax-VL-01`; image-capable config providers auto-register. |
| Audio | OpenAI, Groq, Deepgram, Google, Mistral | Provider transcription (Whisper/Deepgram/Gemini/Voxtral). |
| Video | Google, Qwen, Moonshot | Provider video understanding via vendor plugins; Qwen video understanding uses the Standard DashScope endpoints. |
MiniMax note:

View File

@@ -22,7 +22,7 @@ can still run in OpenClaw, but Anthropic now requires **Extra Usage**
If you want a clearer billing path, use an Anthropic API key instead. OpenClaw
also supports other subscription-style options, including [OpenAI
Codex](/providers/openai), [Alibaba Cloud Model Studio Coding
Codex](/providers/openai), [Qwen Cloud Coding
Plan](/providers/qwen), [MiniMax Coding Plan](/providers/minimax),
and [Z.AI / GLM Coding Plan](/providers/glm).
</Warning>

View File

@@ -73,8 +73,8 @@ Every reply or tool call uses the **current model provider** (OpenAI, Anthropic,
primary source of usage and cost.
This also includes subscription-style hosted providers that still bill outside
OpenClaw's local UI, such as **OpenAI Codex**, **Alibaba Cloud Model Studio
Coding Plan**, **MiniMax Coding Plan**, **Z.AI / GLM Coding Plan**, and
OpenClaw's local UI, such as **OpenAI Codex**, **Qwen Cloud Coding Plan**,
**MiniMax Coding Plan**, **Z.AI / GLM Coding Plan**, and
Anthropic subscription auth with **Extra Usage** enabled.
See [Models](/providers/models) for pricing config and [Token use & costs](/reference/token-use) for display.
@@ -83,9 +83,9 @@ See [Models](/providers/models) for pricing config and [Token use & costs](/refe
Inbound media can be summarized/transcribed before the reply runs. This uses model/provider APIs.
- Audio: OpenAI / Groq / Deepgram (now **auto-enabled** when keys exist).
- Image: OpenAI / Anthropic / Google.
- Video: Google.
- Audio: OpenAI / Groq / Deepgram / Google / Mistral.
- Image: OpenAI / OpenRouter / Anthropic / Google / MiniMax / Moonshot / Qwen / Z.AI.
- Video: Google / Qwen / Moonshot.
See [Media understanding](/nodes/media-understanding).