mirror of
https://github.com/openclaw/openclaw.git
synced 2026-05-08 16:40:42 +00:00
4.1 KiB
4.1 KiB
summary, read_when, title
| summary | read_when | title | ||
|---|---|---|---|---|
| Use NVIDIA's OpenAI-compatible API in OpenClaw |
|
NVIDIA |
NVIDIA provides an OpenAI-compatible API at https://integrate.api.nvidia.com/v1 for
open models for free. Authenticate with an API key from
build.nvidia.com.
Getting started
Create an API key at [build.nvidia.com](https://build.nvidia.com/settings/api-keys). ```bash export NVIDIA_API_KEY="nvapi-..." openclaw onboard --auth-choice nvidia-api-key ``` ```bash openclaw models set nvidia/nvidia/nemotron-3-super-120b-a12b ``` If you pass `--nvidia-api-key` instead of the env var, the value lands in shell history and `ps` output. Prefer the `NVIDIA_API_KEY` environment variable when possible.For non-interactive setup, you can also pass the key directly:
openclaw onboard --auth-choice nvidia-api-key --nvidia-api-key "nvapi-..."
Config example
{
env: { NVIDIA_API_KEY: "nvapi-..." },
models: {
providers: {
nvidia: {
baseUrl: "https://integrate.api.nvidia.com/v1",
api: "openai-completions",
},
},
},
agents: {
defaults: {
model: { primary: "nvidia/nvidia/nemotron-3-super-120b-a12b" },
},
},
}
Built-in catalog
| Model ref | Name | Context | Max output |
|---|---|---|---|
nvidia/nvidia/nemotron-3-super-120b-a12b |
NVIDIA Nemotron 3 Super 120B | 262,144 | 8,192 |
nvidia/moonshotai/kimi-k2.5 |
Kimi K2.5 | 262,144 | 8,192 |
nvidia/minimaxai/minimax-m2.5 |
Minimax M2.5 | 196,608 | 8,192 |
nvidia/z-ai/glm5 |
GLM 5 | 202,752 | 8,192 |
Advanced configuration
The provider auto-enables when the `NVIDIA_API_KEY` environment variable is set. No explicit provider config is required beyond the key. The bundled catalog is static. Costs default to `0` in source since NVIDIA currently offers free API access for the listed models. NVIDIA uses the standard `/v1` completions endpoint. Any OpenAI-compatible tooling should work out of the box with the NVIDIA base URL. Some NVIDIA-hosted custom models can take longer than the default model idle watchdog before they emit a first response chunk. For custom NVIDIA provider entries, raise the provider timeout instead of raising the whole agent runtime timeout:```json5
{
models: {
providers: {
"custom-integrate-api-nvidia-com": {
baseUrl: "https://integrate.api.nvidia.com/v1",
api: "openai-completions",
apiKey: "NVIDIA_API_KEY",
timeoutSeconds: 300,
},
},
},
agents: {
defaults: {
models: {
"custom-integrate-api-nvidia-com/meta/llama-3.1-70b-instruct": {
params: { thinking: "off" },
},
},
},
},
}
```
NVIDIA models are currently free to use. Check
[build.nvidia.com](https://build.nvidia.com/) for the latest availability and
rate-limit details.