Cerebras (docs/providers/cerebras.md): rewrote against extensions/cerebras/openclaw.plugin.json. Added a complete properties summary, CodeGroup for onboarding/direct-flag/env, a Reasoning column on the four-model catalog table (Z.ai GLM 4.7 and GPT OSS 120B are reasoning-capable; Qwen 3 235B and Llama 3.1 8B are not), and a CardGroup of related links. Groq (docs/providers/groq.md): expanded the catalog from 4 hand-picked entries to all 18 bundled models from extensions/groq/openclaw.plugin.json with model refs, reasoning flags, input modalities, and context windows. Removed a stale 'Mixtral 8x7B' row that does not exist in the bundled catalog. Surfaced the audio media-understanding contract (whisper-large-v3-turbo, auto priority 20) as a properties table and explained the per-model reasoning_effort mapping for qwen/qwen3-32b vs the GPT OSS reasoning models. Added an onboarding CodeGroup so the API-key step does not skip 'openclaw onboard --auth-choice groq-api-key'. SGLang (docs/providers/sglang.md): added a properties summary table at the top, including the Qwen/Qwen3-8B model placeholder from extensions/sglang/defaults.ts, the supportsStreamingUsage runtime flag, and the modelPricing.external: false setting. Clarified that the onboarding choice id is bare 'sglang' (custom method) rather than the '-api-key' suffix used by other providers, matching the manifest.
8.1 KiB
summary, title, read_when
| summary | title | read_when | |||
|---|---|---|---|---|---|
| Groq setup (auth + model selection + Whisper transcription) | Groq |
|
Groq provides ultra-fast inference on open-weight models (Llama, Gemma, Kimi, Qwen, GPT OSS, and more) using custom LPU hardware. OpenClaw includes a bundled Groq plugin that registers both an OpenAI-compatible chat provider and an audio media-understanding provider.
| Property | Value |
|---|---|
| Provider id | groq |
| Plugin | bundled, enabledByDefault: true |
| Auth env var | GROQ_API_KEY |
| Onboarding flag | --auth-choice groq-api-key |
| API | OpenAI-compatible (openai-completions) |
| Base URL | https://api.groq.com/openai/v1 |
| Audio transcription | whisper-large-v3-turbo (default) |
| Suggested chat default | groq/llama-3.3-70b-versatile |
Getting started
Create an API key at [console.groq.com/keys](https://console.groq.com/keys).openclaw onboard --auth-choice groq-api-key
export GROQ_API_KEY=gsk_...
</CodeGroup>
```json5
{
agents: {
defaults: {
model: { primary: "groq/llama-3.3-70b-versatile" },
},
},
}
```
```bash
openclaw models list --provider groq
```
Config file example
{
env: { GROQ_API_KEY: "gsk_..." },
agents: {
defaults: {
model: { primary: "groq/llama-3.3-70b-versatile" },
},
},
}
Built-in catalog
OpenClaw ships a manifest-backed Groq catalog with both reasoning and non-reasoning entries. Run openclaw models list --provider groq to see the bundled rows for your installed version, or check console.groq.com/docs/models for Groq's authoritative list.
| Model ref | Name | Reasoning | Input | Context |
|---|---|---|---|---|
groq/llama-3.3-70b-versatile |
Llama 3.3 70B Versatile | no | text | 131,072 |
groq/llama-3.1-8b-instant |
Llama 3.1 8B Instant | no | text | 131,072 |
groq/meta-llama/llama-4-maverick-17b-128e-instruct |
Llama 4 Maverick 17B | no | text + image | 131,072 |
groq/meta-llama/llama-4-scout-17b-16e-instruct |
Llama 4 Scout 17B | no | text + image | 131,072 |
groq/llama3-70b-8192 |
Llama 3 70B | no | text | 8,192 |
groq/llama3-8b-8192 |
Llama 3 8B | no | text | 8,192 |
groq/gemma2-9b-it |
Gemma 2 9B | no | text | 8,192 |
groq/mistral-saba-24b |
Mistral Saba 24B | no | text | 32,768 |
groq/moonshotai/kimi-k2-instruct |
Kimi K2 Instruct | no | text | 131,072 |
groq/moonshotai/kimi-k2-instruct-0905 |
Kimi K2 Instruct 0905 | no | text | 262,144 |
groq/openai/gpt-oss-120b |
GPT OSS 120B | yes | text | 131,072 |
groq/openai/gpt-oss-20b |
GPT OSS 20B | yes | text | 131,072 |
groq/openai/gpt-oss-safeguard-20b |
Safety GPT OSS 20B | yes | text | 131,072 |
groq/qwen-qwq-32b |
Qwen QwQ 32B | yes | text | 131,072 |
groq/qwen/qwen3-32b |
Qwen3 32B | yes | text | 131,072 |
groq/deepseek-r1-distill-llama-70b |
DeepSeek R1 Distill Llama 70B | yes | text | 131,072 |
groq/groq/compound |
Compound | yes | text | 131,072 |
groq/groq/compound-mini |
Compound Mini | yes | text | 131,072 |
Reasoning models
OpenClaw maps its shared /think levels to Groq's model-specific reasoning_effort values:
- For
qwen/qwen3-32b, disabled thinking sendsnoneand enabled thinking sendsdefault. - For Groq GPT OSS reasoning models (
openai/gpt-oss-*), OpenClaw sendslow,medium, orhighbased on/thinklevel. Disabled thinking omitsreasoning_effortbecause those models do not support a disabled value. - DeepSeek R1 Distill, Qwen QwQ, and Compound use Groq's native reasoning surface;
/thinkcontrols visibility but the model always reasons.
See Thinking modes for the shared /think levels and how OpenClaw translates them per provider.
Audio transcription
Groq's bundled plugin also registers an audio media-understanding provider so voice messages can be transcribed through the shared tools.media.audio surface.
| Property | Value |
|---|---|
| Shared config path | tools.media.audio |
| Default base URL | https://api.groq.com/openai/v1 |
| Default model | whisper-large-v3-turbo |
| Auto priority | 20 |
| API endpoint | OpenAI-compatible /audio/transcriptions |
To make Groq the default audio backend:
{
tools: {
media: {
audio: {
models: [{ provider: "groq" }],
},
},
},
}
<Warning>
A key sitting only in `~/.profile` will not help a launchd or systemd daemon unless that environment is imported there too. Set the key in `~/.openclaw/.env` or via `env.shellEnv` to make it readable from the gateway process.
</Warning>
OpenClaw accepts any Groq model id at runtime. Use the exact id shown by Groq and prefix it with `groq/`. The bundled catalog covers the common cases; uncatalogued ids fall through to the default OpenAI-compatible template.
```json5
{
agents: {
defaults: {
model: { primary: "groq/<your-model-id>" },
},
},
}
```