Files
openclaw/docs/providers/groq.md
Vincent Koc 58c706451e docs(providers): rewrite Cerebras, Groq, and SGLang with code-verified setup
Cerebras (docs/providers/cerebras.md): rewrote against
extensions/cerebras/openclaw.plugin.json. Added a complete properties
summary, CodeGroup for onboarding/direct-flag/env, a Reasoning column on
the four-model catalog table (Z.ai GLM 4.7 and GPT OSS 120B are
reasoning-capable; Qwen 3 235B and Llama 3.1 8B are not), and a
CardGroup of related links.

Groq (docs/providers/groq.md): expanded the catalog from 4 hand-picked
entries to all 18 bundled models from extensions/groq/openclaw.plugin.json
with model refs, reasoning flags, input modalities, and context windows.
Removed a stale 'Mixtral 8x7B' row that does not exist in the bundled
catalog. Surfaced the audio media-understanding contract (whisper-large-v3-turbo,
auto priority 20) as a properties table and explained the per-model
reasoning_effort mapping for qwen/qwen3-32b vs the GPT OSS reasoning
models. Added an onboarding CodeGroup so the API-key step does not skip
'openclaw onboard --auth-choice groq-api-key'.

SGLang (docs/providers/sglang.md): added a properties summary table at
the top, including the Qwen/Qwen3-8B model placeholder from
extensions/sglang/defaults.ts, the supportsStreamingUsage runtime flag,
and the modelPricing.external: false setting. Clarified that the
onboarding choice id is bare 'sglang' (custom method) rather than the
'-api-key' suffix used by other providers, matching the manifest.
2026-05-05 16:58:01 -07:00

8.1 KiB

summary, title, read_when
summary title read_when
Groq setup (auth + model selection + Whisper transcription) Groq
You want to use Groq with OpenClaw
You need the API key env var or CLI auth choice
You are configuring Whisper audio transcription on Groq

Groq provides ultra-fast inference on open-weight models (Llama, Gemma, Kimi, Qwen, GPT OSS, and more) using custom LPU hardware. OpenClaw includes a bundled Groq plugin that registers both an OpenAI-compatible chat provider and an audio media-understanding provider.

Property Value
Provider id groq
Plugin bundled, enabledByDefault: true
Auth env var GROQ_API_KEY
Onboarding flag --auth-choice groq-api-key
API OpenAI-compatible (openai-completions)
Base URL https://api.groq.com/openai/v1
Audio transcription whisper-large-v3-turbo (default)
Suggested chat default groq/llama-3.3-70b-versatile

Getting started

Create an API key at [console.groq.com/keys](https://console.groq.com/keys).
openclaw onboard --auth-choice groq-api-key
export GROQ_API_KEY=gsk_...
</CodeGroup>
```json5 { agents: { defaults: { model: { primary: "groq/llama-3.3-70b-versatile" }, }, }, } ``` ```bash openclaw models list --provider groq ```

Config file example

{
  env: { GROQ_API_KEY: "gsk_..." },
  agents: {
    defaults: {
      model: { primary: "groq/llama-3.3-70b-versatile" },
    },
  },
}

Built-in catalog

OpenClaw ships a manifest-backed Groq catalog with both reasoning and non-reasoning entries. Run openclaw models list --provider groq to see the bundled rows for your installed version, or check console.groq.com/docs/models for Groq's authoritative list.

Model ref Name Reasoning Input Context
groq/llama-3.3-70b-versatile Llama 3.3 70B Versatile no text 131,072
groq/llama-3.1-8b-instant Llama 3.1 8B Instant no text 131,072
groq/meta-llama/llama-4-maverick-17b-128e-instruct Llama 4 Maverick 17B no text + image 131,072
groq/meta-llama/llama-4-scout-17b-16e-instruct Llama 4 Scout 17B no text + image 131,072
groq/llama3-70b-8192 Llama 3 70B no text 8,192
groq/llama3-8b-8192 Llama 3 8B no text 8,192
groq/gemma2-9b-it Gemma 2 9B no text 8,192
groq/mistral-saba-24b Mistral Saba 24B no text 32,768
groq/moonshotai/kimi-k2-instruct Kimi K2 Instruct no text 131,072
groq/moonshotai/kimi-k2-instruct-0905 Kimi K2 Instruct 0905 no text 262,144
groq/openai/gpt-oss-120b GPT OSS 120B yes text 131,072
groq/openai/gpt-oss-20b GPT OSS 20B yes text 131,072
groq/openai/gpt-oss-safeguard-20b Safety GPT OSS 20B yes text 131,072
groq/qwen-qwq-32b Qwen QwQ 32B yes text 131,072
groq/qwen/qwen3-32b Qwen3 32B yes text 131,072
groq/deepseek-r1-distill-llama-70b DeepSeek R1 Distill Llama 70B yes text 131,072
groq/groq/compound Compound yes text 131,072
groq/groq/compound-mini Compound Mini yes text 131,072
The catalog evolves with each OpenClaw release. `openclaw models list --provider groq` shows the rows known to your installed version; cross-check with [console.groq.com/docs/models](https://console.groq.com/docs/models) for newly-added or deprecated models.

Reasoning models

OpenClaw maps its shared /think levels to Groq's model-specific reasoning_effort values:

  • For qwen/qwen3-32b, disabled thinking sends none and enabled thinking sends default.
  • For Groq GPT OSS reasoning models (openai/gpt-oss-*), OpenClaw sends low, medium, or high based on /think level. Disabled thinking omits reasoning_effort because those models do not support a disabled value.
  • DeepSeek R1 Distill, Qwen QwQ, and Compound use Groq's native reasoning surface; /think controls visibility but the model always reasons.

See Thinking modes for the shared /think levels and how OpenClaw translates them per provider.

Audio transcription

Groq's bundled plugin also registers an audio media-understanding provider so voice messages can be transcribed through the shared tools.media.audio surface.

Property Value
Shared config path tools.media.audio
Default base URL https://api.groq.com/openai/v1
Default model whisper-large-v3-turbo
Auto priority 20
API endpoint OpenAI-compatible /audio/transcriptions

To make Groq the default audio backend:

{
  tools: {
    media: {
      audio: {
        models: [{ provider: "groq" }],
      },
    },
  },
}
If the Gateway runs as a managed service (launchd, systemd, Docker), `GROQ_API_KEY` must be visible to that process — not just to your interactive shell.
<Warning>
  A key sitting only in `~/.profile` will not help a launchd or systemd daemon unless that environment is imported there too. Set the key in `~/.openclaw/.env` or via `env.shellEnv` to make it readable from the gateway process.
</Warning>
OpenClaw accepts any Groq model id at runtime. Use the exact id shown by Groq and prefix it with `groq/`. The bundled catalog covers the common cases; uncatalogued ids fall through to the default OpenAI-compatible template.
```json5
{
  agents: {
    defaults: {
      model: { primary: "groq/<your-model-id>" },
    },
  },
}
```
Choosing providers, model refs, and failover behavior. Reasoning effort levels and provider-policy interaction. Full config schema including provider and audio settings. Groq dashboard, API docs, and pricing.