mirror of
https://github.com/openclaw/openclaw.git
synced 2026-05-08 14:10:44 +00:00
* feat(nvidia): add NVIDIA provider with onboarding flow Add the NVIDIA build.nvidia.com API as a bundled provider. Default model is nvidia/nvidia/nemotron-3-super-120b-a12b: first segment is the provider id, remaining "nvidia/nemotron-3-super-120b-a12b" is the literal upstream model id (which happens to start with "nvidia/" because NVIDIA is also the model maker). Supporting core change: introduce a provider capability flag nativeIdsIncludeProviderPrefix so providers whose native catalog ids intentionally include their provider prefix (OpenRouter) opt into self-prefix dedupe in modelKey, without hardcoding provider names in core. Providers whose ids merely happen to start with their own name (NVIDIA) leave the flag unset and get the full <provider>/<model-id> concatenation. - extensions/nvidia/*: new plugin, catalog, onboarding, tests, docs - extensions/openrouter/index.ts: declare nativeIdsIncludeProviderPrefix - src/plugins/types.ts: add field to ProviderPlugin - src/plugins/registry.ts: populate self-prefix set on registration - src/agents/provider-self-prefix.ts: sync accessor used by modelKey - src/agents/model-ref-shared.ts: modelKey consults the flag - test updates for affected surfaces Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com> * refactor(model-picker): simplify literal-prefix display to label-only * fix(model-picker): pass workspaceDir/env to allowlist literal-prefix resolution * chore: untrack generated baseline JSON artifacts (gitignored) * fix(nvidia): show literal model ref in picker and onboarding notes * fix(nvidia): show hint whenever display label differs from stored config * fix(nvidia): drop redundant hint from Keep current label * fix(nvidia): restore literal double-prefix display labels * fix(picker): handle literal-prefix fast path * fix(picker): show literal keep label * fix(docs): update nvidia provider docs * fix(nvidia): update test helper imports * fix(changelog): add nvidia provider entry --------- Co-authored-by: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
109 lines
3.3 KiB
Markdown
109 lines
3.3 KiB
Markdown
---
|
|
summary: "Use NVIDIA's OpenAI-compatible API in OpenClaw"
|
|
read_when:
|
|
- You want to use open models in OpenClaw for free
|
|
- You need NVIDIA_API_KEY setup
|
|
title: "NVIDIA"
|
|
---
|
|
|
|
NVIDIA provides an OpenAI-compatible API at `https://integrate.api.nvidia.com/v1` for
|
|
open models for free. Authenticate with an API key from
|
|
[build.nvidia.com](https://build.nvidia.com/settings/api-keys).
|
|
|
|
## Getting started
|
|
|
|
<Steps>
|
|
<Step title="Get your API key">
|
|
Create an API key at [build.nvidia.com](https://build.nvidia.com/settings/api-keys).
|
|
</Step>
|
|
<Step title="Export the key and run onboarding">
|
|
```bash
|
|
export NVIDIA_API_KEY="nvapi-..."
|
|
openclaw onboard --auth-choice nvidia-api-key
|
|
```
|
|
</Step>
|
|
<Step title="Set an NVIDIA model">
|
|
```bash
|
|
openclaw models set nvidia/nvidia/nemotron-3-super-120b-a12b
|
|
```
|
|
</Step>
|
|
</Steps>
|
|
|
|
<Warning>
|
|
If you pass `--nvidia-api-key` instead of the env var, the value lands in shell
|
|
history and `ps` output. Prefer the `NVIDIA_API_KEY` environment variable when
|
|
possible.
|
|
</Warning>
|
|
|
|
For non-interactive setup, you can also pass the key directly:
|
|
|
|
```bash
|
|
openclaw onboard --auth-choice nvidia-api-key --nvidia-api-key "nvapi-..."
|
|
```
|
|
|
|
## Config example
|
|
|
|
```json5
|
|
{
|
|
env: { NVIDIA_API_KEY: "nvapi-..." },
|
|
models: {
|
|
providers: {
|
|
nvidia: {
|
|
baseUrl: "https://integrate.api.nvidia.com/v1",
|
|
api: "openai-completions",
|
|
},
|
|
},
|
|
},
|
|
agents: {
|
|
defaults: {
|
|
model: { primary: "nvidia/nvidia/nemotron-3-super-120b-a12b" },
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
## Built-in catalog
|
|
|
|
| Model ref | Name | Context | Max output |
|
|
| ------------------------------------------ | ---------------------------- | ------- | ---------- |
|
|
| `nvidia/nvidia/nemotron-3-super-120b-a12b` | NVIDIA Nemotron 3 Super 120B | 262,144 | 8,192 |
|
|
| `nvidia/moonshotai/kimi-k2.5` | Kimi K2.5 | 262,144 | 8,192 |
|
|
| `nvidia/minimaxai/minimax-m2.5` | Minimax M2.5 | 196,608 | 8,192 |
|
|
| `nvidia/z-ai/glm5` | GLM 5 | 202,752 | 8,192 |
|
|
|
|
## Advanced configuration
|
|
|
|
<AccordionGroup>
|
|
<Accordion title="Auto-enable behavior">
|
|
The provider auto-enables when the `NVIDIA_API_KEY` environment variable is set.
|
|
No explicit provider config is required beyond the key.
|
|
</Accordion>
|
|
|
|
<Accordion title="Catalog and pricing">
|
|
The bundled catalog is static. Costs default to `0` in source since NVIDIA
|
|
currently offers free API access for the listed models.
|
|
</Accordion>
|
|
|
|
<Accordion title="OpenAI-compatible endpoint">
|
|
NVIDIA uses the standard `/v1` completions endpoint. Any OpenAI-compatible
|
|
tooling should work out of the box with the NVIDIA base URL.
|
|
</Accordion>
|
|
</AccordionGroup>
|
|
|
|
<Tip>
|
|
NVIDIA models are currently free to use. Check
|
|
[build.nvidia.com](https://build.nvidia.com/) for the latest availability and
|
|
rate-limit details.
|
|
</Tip>
|
|
|
|
## Related
|
|
|
|
<CardGroup cols={2}>
|
|
<Card title="Model selection" href="/concepts/model-providers" icon="layers">
|
|
Choosing providers, model refs, and failover behavior.
|
|
</Card>
|
|
<Card title="Configuration reference" href="/gateway/configuration-reference" icon="gear">
|
|
Full config reference for agents, models, and providers.
|
|
</Card>
|
|
</CardGroup>
|