mirror of
https://github.com/openclaw/openclaw.git
synced 2026-05-06 13:00:44 +00:00
Cerebras (docs/providers/cerebras.md): rewrote against extensions/cerebras/openclaw.plugin.json. Added a complete properties summary, CodeGroup for onboarding/direct-flag/env, a Reasoning column on the four-model catalog table (Z.ai GLM 4.7 and GPT OSS 120B are reasoning-capable; Qwen 3 235B and Llama 3.1 8B are not), and a CardGroup of related links. Groq (docs/providers/groq.md): expanded the catalog from 4 hand-picked entries to all 18 bundled models from extensions/groq/openclaw.plugin.json with model refs, reasoning flags, input modalities, and context windows. Removed a stale 'Mixtral 8x7B' row that does not exist in the bundled catalog. Surfaced the audio media-understanding contract (whisper-large-v3-turbo, auto priority 20) as a properties table and explained the per-model reasoning_effort mapping for qwen/qwen3-32b vs the GPT OSS reasoning models. Added an onboarding CodeGroup so the API-key step does not skip 'openclaw onboard --auth-choice groq-api-key'. SGLang (docs/providers/sglang.md): added a properties summary table at the top, including the Qwen/Qwen3-8B model placeholder from extensions/sglang/defaults.ts, the supportsStreamingUsage runtime flag, and the modelPricing.external: false setting. Clarified that the onboarding choice id is bare 'sglang' (custom method) rather than the '-api-key' suffix used by other providers, matching the manifest.
160 lines
5.0 KiB
Markdown
160 lines
5.0 KiB
Markdown
---
|
|
summary: "Run OpenClaw with SGLang (OpenAI-compatible self-hosted server)"
|
|
read_when:
|
|
- You want to run OpenClaw against a local SGLang server
|
|
- You want OpenAI-compatible /v1 endpoints with your own models
|
|
title: "SGLang"
|
|
---
|
|
|
|
SGLang serves open-weight models via an OpenAI-compatible HTTP API. OpenClaw connects to SGLang using the `openai-completions` provider family with auto-discovery of available models.
|
|
|
|
| Property | Value |
|
|
| ------------------------- | ------------------------------------------------------------ |
|
|
| Provider id | `sglang` |
|
|
| Plugin | bundled, `enabledByDefault: true` |
|
|
| Auth env var | `SGLANG_API_KEY` (any non-empty value if server has no auth) |
|
|
| Onboarding flag | `--auth-choice sglang` |
|
|
| API | OpenAI-compatible (`openai-completions`) |
|
|
| Default base URL | `http://127.0.0.1:30000/v1` |
|
|
| Default model placeholder | `sglang/Qwen/Qwen3-8B` |
|
|
| Streaming usage | Yes (`supportsStreamingUsage: true`) |
|
|
| Pricing | Marked external-free (`modelPricing.external: false`) |
|
|
|
|
OpenClaw also **auto-discovers** available models from SGLang when you opt in with `SGLANG_API_KEY` and you do not define an explicit `models.providers.sglang` entry — see [Model discovery (implicit provider)](#model-discovery-implicit-provider) below.
|
|
|
|
## Getting started
|
|
|
|
<Steps>
|
|
<Step title="Start SGLang">
|
|
Launch SGLang with an OpenAI-compatible server. Your base URL should expose
|
|
`/v1` endpoints (for example `/v1/models`, `/v1/chat/completions`). SGLang
|
|
commonly runs on:
|
|
|
|
- `http://127.0.0.1:30000/v1`
|
|
|
|
</Step>
|
|
<Step title="Set an API key">
|
|
Any value works if no auth is configured on your server:
|
|
|
|
```bash
|
|
export SGLANG_API_KEY="sglang-local"
|
|
```
|
|
|
|
</Step>
|
|
<Step title="Run onboarding or set a model directly">
|
|
```bash
|
|
openclaw onboard
|
|
```
|
|
|
|
Or configure the model manually:
|
|
|
|
```json5
|
|
{
|
|
agents: {
|
|
defaults: {
|
|
model: { primary: "sglang/your-model-id" },
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
</Step>
|
|
</Steps>
|
|
|
|
## Model discovery (implicit provider)
|
|
|
|
When `SGLANG_API_KEY` is set (or an auth profile exists) and you **do not**
|
|
define `models.providers.sglang`, OpenClaw will query:
|
|
|
|
- `GET http://127.0.0.1:30000/v1/models`
|
|
|
|
and convert the returned IDs into model entries.
|
|
|
|
<Note>
|
|
If you set `models.providers.sglang` explicitly, auto-discovery is skipped and
|
|
you must define models manually.
|
|
</Note>
|
|
|
|
## Explicit configuration (manual models)
|
|
|
|
Use explicit config when:
|
|
|
|
- SGLang runs on a different host/port.
|
|
- You want to pin `contextWindow`/`maxTokens` values.
|
|
- Your server requires a real API key (or you want to control headers).
|
|
|
|
```json5
|
|
{
|
|
models: {
|
|
providers: {
|
|
sglang: {
|
|
baseUrl: "http://127.0.0.1:30000/v1",
|
|
apiKey: "${SGLANG_API_KEY}",
|
|
api: "openai-completions",
|
|
models: [
|
|
{
|
|
id: "your-model-id",
|
|
name: "Local SGLang Model",
|
|
reasoning: false,
|
|
input: ["text"],
|
|
cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 },
|
|
contextWindow: 128000,
|
|
maxTokens: 8192,
|
|
},
|
|
],
|
|
},
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
## Advanced configuration
|
|
|
|
<AccordionGroup>
|
|
<Accordion title="Proxy-style behavior">
|
|
SGLang is treated as a proxy-style OpenAI-compatible `/v1` backend, not a
|
|
native OpenAI endpoint.
|
|
|
|
| Behavior | SGLang |
|
|
|----------|--------|
|
|
| OpenAI-only request shaping | Not applied |
|
|
| `service_tier`, Responses `store`, prompt-cache hints | Not sent |
|
|
| Reasoning-compat payload shaping | Not applied |
|
|
| Hidden attribution headers (`originator`, `version`, `User-Agent`) | Not injected on custom SGLang base URLs |
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="Troubleshooting">
|
|
**Server not reachable**
|
|
|
|
Verify the server is running and responding:
|
|
|
|
```bash
|
|
curl http://127.0.0.1:30000/v1/models
|
|
```
|
|
|
|
**Auth errors**
|
|
|
|
If requests fail with auth errors, set a real `SGLANG_API_KEY` that matches
|
|
your server configuration, or configure the provider explicitly under
|
|
`models.providers.sglang`.
|
|
|
|
<Tip>
|
|
If you run SGLang without authentication, any non-empty value for
|
|
`SGLANG_API_KEY` is sufficient to opt in to model discovery.
|
|
</Tip>
|
|
|
|
</Accordion>
|
|
</AccordionGroup>
|
|
|
|
## Related
|
|
|
|
<CardGroup cols={2}>
|
|
<Card title="Model selection" href="/concepts/model-providers" icon="layers">
|
|
Choosing providers, model refs, and failover behavior.
|
|
</Card>
|
|
<Card title="Configuration reference" href="/gateway/configuration-reference" icon="gear">
|
|
Full config schema including provider entries.
|
|
</Card>
|
|
</CardGroup>
|