refactor(nvidia-endpoints): updated language & default models (#59866)

* fix(nvidia-endpoints): updated language & default models

* fix(nvidia-endpoints): updated link for api key

* fix(nvidia-endpoints): removed unused const

* fix(nvidia-endpoints): edited max tokens

* fix(nvidia-endpoints): fixed typo

---------

Co-authored-by: Devin Robison <drobison00@users.noreply.github.com>
This commit is contained in:
nv-kasikritc
2026-04-07 21:47:29 +07:00
committed by GitHub
parent ac6693986b
commit d43cc470c6
3 changed files with 36 additions and 26 deletions

View File

@@ -1,14 +1,14 @@
---
summary: "Use NVIDIA's OpenAI-compatible API in OpenClaw"
read_when:
- You want to use NVIDIA models in OpenClaw
- You want to use open models in OpenClaw for free
- You need NVIDIA_API_KEY setup
title: "NVIDIA"
---
# NVIDIA
NVIDIA provides an OpenAI-compatible API at `https://integrate.api.nvidia.com/v1` for Nemotron and NeMo models. Authenticate with an API key from [NVIDIA NGC](https://catalog.ngc.nvidia.com/).
NVIDIA provides an OpenAI-compatible API at `https://integrate.api.nvidia.com/v1` for open models for free. Authenticate with an API key from [build.nvidia.com](https://build.nvidia.com/settings/api-keys).
## CLI setup
@@ -17,7 +17,7 @@ Export the key once, then run onboarding and set an NVIDIA model:
```bash
export NVIDIA_API_KEY="nvapi-..."
openclaw onboard --auth-choice skip
openclaw models set nvidia/nvidia/llama-3.1-nemotron-70b-instruct
openclaw models set nvidia/nvidia/nemotron-3-super-120b-a12b
```
If you still pass `--token`, remember it lands in shell history and `ps` output; prefer the env var when possible.
@@ -37,7 +37,7 @@ If you still pass `--token`, remember it lands in shell history and `ps` output;
},
agents: {
defaults: {
model: { primary: "nvidia/nvidia/llama-3.1-nemotron-70b-instruct" },
model: { primary: "nvidia/nvidia/nemotron-3-super-120b-a12b" },
},
},
}
@@ -45,14 +45,15 @@ If you still pass `--token`, remember it lands in shell history and `ps` output;
## Model IDs
| Model ref | Name | Context | Max output |
| ---------------------------------------------------- | ---------------------------------------- | ------- | ---------- |
| `nvidia/nvidia/llama-3.1-nemotron-70b-instruct` | NVIDIA Llama 3.1 Nemotron 70B Instruct | 131,072 | 4,096 |
| `nvidia/meta/llama-3.3-70b-instruct` | Meta Llama 3.3 70B Instruct | 131,072 | 4,096 |
| `nvidia/nvidia/mistral-nemo-minitron-8b-8k-instruct` | NVIDIA Mistral NeMo Minitron 8B Instruct | 8,192 | 2,048 |
| Model ref | Name | Context | Max output |
| ------------------------------------------ | ---------------------------- | ------- | ---------- |
| `nvidia/nvidia/nemotron-3-super-120b-a12b` | NVIDIA Nemotron 3 Super 120B | 262,144 | 8,192 |
| `nvidia/moonshotai/kimi-k2.5` | Kimi K2.5 | 262,144 | 8,192 |
| `nvidia/minimaxai/minimax-m2.5` | Minimax M2.5 | 196,608 | 8,192 |
| `nvidia/z-ai/glm5` | GLM 5 | 202,752 | 8,192 |
## Notes
- OpenAI-compatible `/v1` endpoint; use an API key from NVIDIA NGC.
- OpenAI-compatible `/v1` endpoint; use an API key from [build.nvidia.com](https://build.nvidia.com/).
- Provider auto-enables when `NVIDIA_API_KEY` is set.
- The bundled catalog is static; costs default to `0` in source.