mirror of
https://github.com/openclaw/openclaw.git
synced 2026-05-08 13:30:43 +00:00
* feat(litellm): add image generation provider Registers litellm as an image-generation provider so model refs like litellm/gpt-image-2 route through the LiteLLM proxy, and agents.defaults.imageGenerationModel.fallbacks entries of the form litellm/... resolve without "No image-generation provider registered for litellm" errors. Implementation uses the OpenAI-compatible /images/generations and /images/edits endpoints that LiteLLM proxies for. BaseUrl resolves from models.providers.litellm.baseUrl (default http://localhost:4000). Private network is auto-allowed when baseUrl is a loopback/RFC1918 address, which covers the common self-hosted LiteLLM proxy case without needing OPENCLAW_PROVIDER_ALLOW_PRIVATE_NETWORK. Public baseUrls keep normal SSRF defaults. Default model is gpt-image-2 (matching upstream 4.21+ OpenAI default). Advertises the same 2K/4K sizes OpenAI now exposes, plus legacy 256/512/1024 for dall-e-3. Supports both generate and edit. Local patch. LiteLLM has no upstream image-generation support yet; revisit if upstream adds one. * ci: rerun after upstream main hot-fix * fix(litellm): harden image generation provider --------- Co-authored-by: Chris Zhang <chris@ChrisdeMac-mini.local> Co-authored-by: Peter Steinberger <steipete@gmail.com>
229 lines
5.9 KiB
Markdown
229 lines
5.9 KiB
Markdown
---
|
|
summary: "Run OpenClaw through LiteLLM Proxy for unified model access and cost tracking"
|
|
title: "LiteLLM"
|
|
read_when:
|
|
- You want to route OpenClaw through a LiteLLM proxy
|
|
- You need cost tracking, logging, or model routing through LiteLLM
|
|
---
|
|
|
|
[LiteLLM](https://litellm.ai) is an open-source LLM gateway that provides a unified API to 100+ model providers. Route OpenClaw through LiteLLM to get centralized cost tracking, logging, and the flexibility to switch backends without changing your OpenClaw config.
|
|
|
|
<Tip>
|
|
**Why use LiteLLM with OpenClaw?**
|
|
|
|
- **Cost tracking** — See exactly what OpenClaw spends across all models
|
|
- **Model routing** — Switch between Claude, GPT-4, Gemini, Bedrock without config changes
|
|
- **Virtual keys** — Create keys with spend limits for OpenClaw
|
|
- **Logging** — Full request/response logs for debugging
|
|
- **Fallbacks** — Automatic failover if your primary provider is down
|
|
|
|
</Tip>
|
|
|
|
## Quick start
|
|
|
|
<Tabs>
|
|
<Tab title="Onboarding (recommended)">
|
|
**Best for:** fastest path to a working LiteLLM setup.
|
|
|
|
<Steps>
|
|
<Step title="Run onboarding">
|
|
```bash
|
|
openclaw onboard --auth-choice litellm-api-key
|
|
```
|
|
</Step>
|
|
</Steps>
|
|
|
|
</Tab>
|
|
|
|
<Tab title="Manual setup">
|
|
**Best for:** full control over installation and config.
|
|
|
|
<Steps>
|
|
<Step title="Start LiteLLM Proxy">
|
|
```bash
|
|
pip install 'litellm[proxy]'
|
|
litellm --model claude-opus-4-6
|
|
```
|
|
</Step>
|
|
<Step title="Point OpenClaw to LiteLLM">
|
|
```bash
|
|
export LITELLM_API_KEY="your-litellm-key"
|
|
|
|
openclaw
|
|
```
|
|
|
|
That's it. OpenClaw now routes through LiteLLM.
|
|
</Step>
|
|
</Steps>
|
|
|
|
</Tab>
|
|
</Tabs>
|
|
|
|
## Configuration
|
|
|
|
### Environment variables
|
|
|
|
```bash
|
|
export LITELLM_API_KEY="sk-litellm-key"
|
|
```
|
|
|
|
### Config file
|
|
|
|
```json5
|
|
{
|
|
models: {
|
|
providers: {
|
|
litellm: {
|
|
baseUrl: "http://localhost:4000",
|
|
apiKey: "${LITELLM_API_KEY}",
|
|
api: "openai-completions",
|
|
models: [
|
|
{
|
|
id: "claude-opus-4-6",
|
|
name: "Claude Opus 4.6",
|
|
reasoning: true,
|
|
input: ["text", "image"],
|
|
contextWindow: 200000,
|
|
maxTokens: 64000,
|
|
},
|
|
{
|
|
id: "gpt-4o",
|
|
name: "GPT-4o",
|
|
reasoning: false,
|
|
input: ["text", "image"],
|
|
contextWindow: 128000,
|
|
maxTokens: 8192,
|
|
},
|
|
],
|
|
},
|
|
},
|
|
},
|
|
agents: {
|
|
defaults: {
|
|
model: { primary: "litellm/claude-opus-4-6" },
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
## Advanced configuration
|
|
|
|
### Image generation
|
|
|
|
LiteLLM can also back the `image_generate` tool through OpenAI-compatible
|
|
`/images/generations` and `/images/edits` routes. Configure a LiteLLM image
|
|
model under `agents.defaults.imageGenerationModel`:
|
|
|
|
```json5
|
|
{
|
|
models: {
|
|
providers: {
|
|
litellm: {
|
|
baseUrl: "http://localhost:4000",
|
|
apiKey: "${LITELLM_API_KEY}",
|
|
},
|
|
},
|
|
},
|
|
agents: {
|
|
defaults: {
|
|
imageGenerationModel: {
|
|
primary: "litellm/gpt-image-2",
|
|
timeoutMs: 180_000,
|
|
},
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
Loopback LiteLLM URLs such as `http://localhost:4000` work without a global
|
|
private-network override. For a LAN-hosted proxy, set
|
|
`models.providers.litellm.request.allowPrivateNetwork: true` because the API key
|
|
will be sent to the configured proxy host.
|
|
|
|
<AccordionGroup>
|
|
<Accordion title="Virtual keys">
|
|
Create a dedicated key for OpenClaw with spend limits:
|
|
|
|
```bash
|
|
curl -X POST "http://localhost:4000/key/generate" \
|
|
-H "Authorization: Bearer $LITELLM_MASTER_KEY" \
|
|
-H "Content-Type: application/json" \
|
|
-d '{
|
|
"key_alias": "openclaw",
|
|
"max_budget": 50.00,
|
|
"budget_duration": "monthly"
|
|
}'
|
|
```
|
|
|
|
Use the generated key as `LITELLM_API_KEY`.
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="Model routing">
|
|
LiteLLM can route model requests to different backends. Configure in your LiteLLM `config.yaml`:
|
|
|
|
```yaml
|
|
model_list:
|
|
- model_name: claude-opus-4-6
|
|
litellm_params:
|
|
model: claude-opus-4-6
|
|
api_key: os.environ/ANTHROPIC_API_KEY
|
|
|
|
- model_name: gpt-4o
|
|
litellm_params:
|
|
model: gpt-4o
|
|
api_key: os.environ/OPENAI_API_KEY
|
|
```
|
|
|
|
OpenClaw keeps requesting `claude-opus-4-6` — LiteLLM handles the routing.
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="Viewing usage">
|
|
Check LiteLLM's dashboard or API:
|
|
|
|
```bash
|
|
# Key info
|
|
curl "http://localhost:4000/key/info" \
|
|
-H "Authorization: Bearer sk-litellm-key"
|
|
|
|
# Spend logs
|
|
curl "http://localhost:4000/spend/logs" \
|
|
-H "Authorization: Bearer $LITELLM_MASTER_KEY"
|
|
```
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="Proxy behavior notes">
|
|
- LiteLLM runs on `http://localhost:4000` by default
|
|
- OpenClaw connects through LiteLLM's proxy-style OpenAI-compatible `/v1`
|
|
endpoint
|
|
- Native OpenAI-only request shaping does not apply through LiteLLM:
|
|
no `service_tier`, no Responses `store`, no prompt-cache hints, and no
|
|
OpenAI reasoning-compat payload shaping
|
|
- Hidden OpenClaw attribution headers (`originator`, `version`, `User-Agent`)
|
|
are not injected on custom LiteLLM base URLs
|
|
</Accordion>
|
|
</AccordionGroup>
|
|
|
|
<Note>
|
|
For general provider configuration and failover behavior, see [Model Providers](/concepts/model-providers).
|
|
</Note>
|
|
|
|
## Related
|
|
|
|
<CardGroup cols={2}>
|
|
<Card title="LiteLLM Docs" href="https://docs.litellm.ai" icon="book">
|
|
Official LiteLLM documentation and API reference.
|
|
</Card>
|
|
<Card title="Model selection" href="/concepts/model-providers" icon="layers">
|
|
Overview of all providers, model refs, and failover behavior.
|
|
</Card>
|
|
<Card title="Configuration" href="/gateway/configuration" icon="gear">
|
|
Full config reference.
|
|
</Card>
|
|
<Card title="Model selection" href="/concepts/models" icon="brain">
|
|
How to choose and configure models.
|
|
</Card>
|
|
</CardGroup>
|