Files
openclaw/docs/providers/nvidia.md
Agustin Rivera 9a0b43c47e feat(nvidia): add NVIDIA provider with onboarding flow (#71204)
* feat(nvidia): add NVIDIA provider with onboarding flow

Add the NVIDIA build.nvidia.com API as a bundled provider. Default model
is nvidia/nvidia/nemotron-3-super-120b-a12b: first segment is the provider
id, remaining "nvidia/nemotron-3-super-120b-a12b" is the literal upstream
model id (which happens to start with "nvidia/" because NVIDIA is also the
model maker).

Supporting core change: introduce a provider capability flag
nativeIdsIncludeProviderPrefix so providers whose native catalog ids
intentionally include their provider prefix (OpenRouter) opt into self-prefix
dedupe in modelKey, without hardcoding provider names in core. Providers
whose ids merely happen to start with their own name (NVIDIA) leave the flag
unset and get the full <provider>/<model-id> concatenation.

- extensions/nvidia/*: new plugin, catalog, onboarding, tests, docs
- extensions/openrouter/index.ts: declare nativeIdsIncludeProviderPrefix
- src/plugins/types.ts: add field to ProviderPlugin
- src/plugins/registry.ts: populate self-prefix set on registration
- src/agents/provider-self-prefix.ts: sync accessor used by modelKey
- src/agents/model-ref-shared.ts: modelKey consults the flag
- test updates for affected surfaces

Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>

* refactor(model-picker): simplify literal-prefix display to label-only

* fix(model-picker): pass workspaceDir/env to allowlist literal-prefix resolution

* chore: untrack generated baseline JSON artifacts (gitignored)

* fix(nvidia): show literal model ref in picker and onboarding notes

* fix(nvidia): show hint whenever display label differs from stored config

* fix(nvidia): drop redundant hint from Keep current label

* fix(nvidia): restore literal double-prefix display labels

* fix(picker): handle literal-prefix fast path

* fix(picker): show literal keep label

* fix(docs): update nvidia provider docs

* fix(nvidia): update test helper imports

* fix(changelog): add nvidia provider entry

---------

Co-authored-by: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
2026-04-29 09:25:12 -07:00

3.3 KiB

summary, read_when, title
summary read_when title
Use NVIDIA's OpenAI-compatible API in OpenClaw
You want to use open models in OpenClaw for free
You need NVIDIA_API_KEY setup
NVIDIA

NVIDIA provides an OpenAI-compatible API at https://integrate.api.nvidia.com/v1 for open models for free. Authenticate with an API key from build.nvidia.com.

Getting started

Create an API key at [build.nvidia.com](https://build.nvidia.com/settings/api-keys). ```bash export NVIDIA_API_KEY="nvapi-..." openclaw onboard --auth-choice nvidia-api-key ``` ```bash openclaw models set nvidia/nvidia/nemotron-3-super-120b-a12b ``` If you pass `--nvidia-api-key` instead of the env var, the value lands in shell history and `ps` output. Prefer the `NVIDIA_API_KEY` environment variable when possible.

For non-interactive setup, you can also pass the key directly:

openclaw onboard --auth-choice nvidia-api-key --nvidia-api-key "nvapi-..."

Config example

{
  env: { NVIDIA_API_KEY: "nvapi-..." },
  models: {
    providers: {
      nvidia: {
        baseUrl: "https://integrate.api.nvidia.com/v1",
        api: "openai-completions",
      },
    },
  },
  agents: {
    defaults: {
      model: { primary: "nvidia/nvidia/nemotron-3-super-120b-a12b" },
    },
  },
}

Built-in catalog

Model ref Name Context Max output
nvidia/nvidia/nemotron-3-super-120b-a12b NVIDIA Nemotron 3 Super 120B 262,144 8,192
nvidia/moonshotai/kimi-k2.5 Kimi K2.5 262,144 8,192
nvidia/minimaxai/minimax-m2.5 Minimax M2.5 196,608 8,192
nvidia/z-ai/glm5 GLM 5 202,752 8,192

Advanced configuration

The provider auto-enables when the `NVIDIA_API_KEY` environment variable is set. No explicit provider config is required beyond the key. The bundled catalog is static. Costs default to `0` in source since NVIDIA currently offers free API access for the listed models. NVIDIA uses the standard `/v1` completions endpoint. Any OpenAI-compatible tooling should work out of the box with the NVIDIA base URL. NVIDIA models are currently free to use. Check [build.nvidia.com](https://build.nvidia.com/) for the latest availability and rate-limit details. Choosing providers, model refs, and failover behavior. Full config reference for agents, models, and providers.