fix: align NVIDIA provider docs and model ids (#11606)

This commit is contained in:
Peter Steinberger
2026-02-14 05:35:40 +01:00
parent e0132514f6
commit fdda261478
3 changed files with 8 additions and 7 deletions

View File

@@ -12,11 +12,12 @@ NVIDIA provides an OpenAI-compatible API at `https://integrate.api.nvidia.com/v1
## CLI setup ## CLI setup
Export the key once, then run onboarding without an inline secret: Export the key once, then run onboarding and set an NVIDIA model:
```bash ```bash
export NVIDIA_API_KEY="nvapi-..." export NVIDIA_API_KEY="nvapi-..."
openclaw onboard --auth-choice apiKey --token-provider nvidia openclaw onboard --auth-choice skip
openclaw models set nvidia/nvidia/llama-3.1-nemotron-70b-instruct
``` ```
If you still pass `--token`, remember it lands in shell history and `ps` output; prefer the env var when possible. If you still pass `--token`, remember it lands in shell history and `ps` output; prefer the env var when possible.
@@ -36,7 +37,7 @@ If you still pass `--token`, remember it lands in shell history and `ps` output;
}, },
agents: { agents: {
defaults: { defaults: {
model: { primary: "nvidia/llama-3.1-nemotron-70b-instruct" }, model: { primary: "nvidia/nvidia/llama-3.1-nemotron-70b-instruct" },
}, },
}, },
} }
@@ -45,7 +46,7 @@ If you still pass `--token`, remember it lands in shell history and `ps` output;
## Model IDs ## Model IDs
- `nvidia/llama-3.1-nemotron-70b-instruct` (default) - `nvidia/llama-3.1-nemotron-70b-instruct` (default)
- `nvidia/llama-3.3-70b-instruct` - `meta/llama-3.3-70b-instruct`
- `nvidia/mistral-nemo-minitron-8b-8k-instruct` - `nvidia/mistral-nemo-minitron-8b-8k-instruct`
## Notes ## Notes

View File

@@ -59,7 +59,7 @@ describe("NVIDIA provider", () => {
const provider = buildNvidiaProvider(); const provider = buildNvidiaProvider();
const modelIds = provider.models.map((m) => m.id); const modelIds = provider.models.map((m) => m.id);
expect(modelIds).toContain("nvidia/llama-3.1-nemotron-70b-instruct"); expect(modelIds).toContain("nvidia/llama-3.1-nemotron-70b-instruct");
expect(modelIds).toContain("nvidia/llama-3.3-70b-instruct"); expect(modelIds).toContain("meta/llama-3.3-70b-instruct");
expect(modelIds).toContain("nvidia/mistral-nemo-minitron-8b-8k-instruct"); expect(modelIds).toContain("nvidia/mistral-nemo-minitron-8b-8k-instruct");
}); });
}); });

View File

@@ -635,8 +635,8 @@ export function buildNvidiaProvider(): ProviderConfig {
maxTokens: NVIDIA_DEFAULT_MAX_TOKENS, maxTokens: NVIDIA_DEFAULT_MAX_TOKENS,
}, },
{ {
id: "nvidia/llama-3.3-70b-instruct", id: "meta/llama-3.3-70b-instruct",
name: "NVIDIA Llama 3.3 70B Instruct", name: "Meta Llama 3.3 70B Instruct",
reasoning: false, reasoning: false,
input: ["text"], input: ["text"],
cost: NVIDIA_DEFAULT_COST, cost: NVIDIA_DEFAULT_COST,