Files
openclaw/docs/providers/mistral.md
Vincent Koc 813fe0a3be docs(providers): rewrite Tencent, Mistral, and Inferrs with code-verified setup
Tencent (docs/providers/tencent.md): rewrote against
extensions/tencent/openclaw.plugin.json. Removed the duplicate
'# Tencent Cloud TokenHub' H1 (Mintlify renders title from frontmatter;
the in-body H1 produces a brittle anchor). Added a properties summary
with onboarding flag and direct CLI flag. Promoted the Quick Start
auth step into a CodeGroup covering onboarding/direct/env. Added a
tiered-pricing table sourced from the bundled cost.tieredPricing
metadata so cost expectations are visible alongside the catalog.
Replaced trailing bullet list under 'Related documentation' with a
CardGroup pointing at TokenHub product/model-card pages.

Mistral (docs/providers/mistral.md): replaced the 3-bullet provider
header with a properties table that surfaces all four contracts the
plugin registers (chat completions via the model catalog, media
understanding Voxtral batch, voice-call streaming Voxtral Realtime, and
memory embeddings via mistral-embed) plus the onboarding flag and
direct CLI flag. Removed a stray 'Z.AI uses Bearer auth' line that had
leaked into the 'Auth and base URL' accordion from a copy-paste
elsewhere; replaced it with a Mistral-correct base-URL override note.
Verified the seven-row LLM catalog matches plugin manifest model ids.

Inferrs (docs/providers/inferrs.md): added a properties table
explicitly stating that inferrs is NOT a bundled plugin and is
configured under models.providers.inferrs (no onboarding choice flag,
no extension package). Cross-linked SGLang and vLLM as bundled
self-hosted alternatives so readers picking a local backend can
compare.
2026-05-05 17:24:53 -07:00

7.5 KiB

summary, read_when, title
summary read_when title
Use Mistral models and Voxtral transcription with OpenClaw
You want to use Mistral models in OpenClaw
You want Voxtral realtime transcription for Voice Call
You need Mistral API key onboarding and model refs
Mistral

OpenClaw includes a bundled Mistral plugin that registers four contracts: chat completions, media understanding (Voxtral batch transcription), realtime STT for Voice Call (Voxtral Realtime), and memory embeddings (mistral-embed).

Property Value
Provider id mistral
Plugin bundled, enabledByDefault: true
Auth env var MISTRAL_API_KEY
Onboarding flag --auth-choice mistral-api-key
Direct CLI flag --mistral-api-key <key>
API OpenAI-compatible (openai-completions)
Base URL https://api.mistral.ai/v1
Default model mistral/mistral-large-latest
Embedding model mistral-embed
Voxtral batch voxtral-mini-latest (audio transcription)
Voxtral realtime voxtral-mini-transcribe-realtime-2602

Getting started

Create an API key in the [Mistral Console](https://console.mistral.ai/). ```bash openclaw onboard --auth-choice mistral-api-key ```
Or pass the key directly:

```bash
openclaw onboard --mistral-api-key "$MISTRAL_API_KEY"
```
```json5 { env: { MISTRAL_API_KEY: "sk-..." }, agents: { defaults: { model: { primary: "mistral/mistral-large-latest" } } }, } ``` ```bash openclaw models list --provider mistral ```

Built-in LLM catalog

OpenClaw currently ships this bundled Mistral catalog:

Model ref Input Context Max output Notes
mistral/mistral-large-latest text, image 262,144 16,384 Default model
mistral/mistral-medium-2508 text, image 262,144 8,192 Mistral Medium 3.1
mistral/mistral-small-latest text, image 128,000 16,384 Mistral Small 4; adjustable reasoning via API reasoning_effort
mistral/pixtral-large-latest text, image 128,000 32,768 Pixtral
mistral/codestral-latest text 256,000 4,096 Coding
mistral/devstral-medium-latest text 262,144 32,768 Devstral 2
mistral/magistral-small text 128,000 40,000 Reasoning-enabled

Audio transcription (Voxtral)

Use Voxtral for batch audio transcription through the media understanding pipeline.

{
  tools: {
    media: {
      audio: {
        enabled: true,
        models: [{ provider: "mistral", model: "voxtral-mini-latest" }],
      },
    },
  },
}
The media transcription path uses `/v1/audio/transcriptions`. The default audio model for Mistral is `voxtral-mini-latest`.

Voice Call streaming STT

The bundled mistral plugin registers Voxtral Realtime as a Voice Call streaming STT provider.

Setting Config path Default
API key plugins.entries.voice-call.config.streaming.providers.mistral.apiKey Falls back to MISTRAL_API_KEY
Model ...mistral.model voxtral-mini-transcribe-realtime-2602
Encoding ...mistral.encoding pcm_mulaw
Sample rate ...mistral.sampleRate 8000
Target delay ...mistral.targetStreamingDelayMs 800
{
  plugins: {
    entries: {
      "voice-call": {
        config: {
          streaming: {
            enabled: true,
            provider: "mistral",
            providers: {
              mistral: {
                apiKey: "${MISTRAL_API_KEY}",
                targetStreamingDelayMs: 800,
              },
            },
          },
        },
      },
    },
  },
}
OpenClaw defaults Mistral realtime STT to `pcm_mulaw` at 8 kHz so Voice Call can forward Twilio media frames directly. Use `encoding: "pcm_s16le"` and a matching `sampleRate` only if your upstream stream is already raw PCM.

Advanced configuration

`mistral/mistral-small-latest` maps to Mistral Small 4 and supports [adjustable reasoning](https://docs.mistral.ai/capabilities/reasoning/adjustable) on the Chat Completions API via `reasoning_effort` (`none` minimizes extra thinking in the output; `high` surfaces full thinking traces before the final answer).
OpenClaw maps the session **thinking** level to Mistral's API:

| OpenClaw thinking level                          | Mistral `reasoning_effort` |
| ------------------------------------------------ | -------------------------- |
| **off** / **minimal**                            | `none`                     |
| **low** / **medium** / **high** / **xhigh** / **adaptive** / **max** | `high`     |

<Note>
Other bundled Mistral catalog models do not use this parameter. Keep using `magistral-*` models when you want Mistral's native reasoning-first behavior.
</Note>
Mistral can serve memory embeddings via `/v1/embeddings` (default model: `mistral-embed`).
```json5
{
  memorySearch: { provider: "mistral" },
}
```
- Mistral auth uses `MISTRAL_API_KEY` (Bearer header). - Provider base URL defaults to `https://api.mistral.ai/v1` and accepts the standard OpenAI-compatible chat-completions request shape. - Onboarding default model is `mistral/mistral-large-latest`. - Override the base URL under `models.providers.mistral.baseUrl` only when Mistral explicitly publishes a regional endpoint you need. Choosing providers, model refs, and failover behavior. Audio transcription setup and provider selection.