mirror of
https://github.com/openclaw/openclaw.git
synced 2026-05-06 14:20:44 +00:00
1011 lines
33 KiB
Markdown
1011 lines
33 KiB
Markdown
---
|
|
summary: "Run OpenClaw with Ollama (cloud and local models)"
|
|
read_when:
|
|
- You want to run OpenClaw with cloud or local models via Ollama
|
|
- You need Ollama setup and configuration guidance
|
|
- You want Ollama vision models for image understanding
|
|
title: "Ollama"
|
|
---
|
|
|
|
OpenClaw integrates with Ollama's native API (`/api/chat`) for hosted cloud models and local/self-hosted Ollama servers. You can use Ollama in three modes: `Cloud + Local` through a reachable Ollama host, `Cloud only` against `https://ollama.com`, or `Local only` against a reachable Ollama host.
|
|
|
|
<Warning>
|
|
**Remote Ollama users**: Do not use the `/v1` OpenAI-compatible URL (`http://host:11434/v1`) with OpenClaw. This breaks tool calling and models may output raw tool JSON as plain text. Use the native Ollama API URL instead: `baseUrl: "http://host:11434"` (no `/v1`).
|
|
</Warning>
|
|
|
|
Ollama provider config uses `baseUrl` as the canonical key. OpenClaw also accepts `baseURL` for compatibility with OpenAI SDK-style examples, but new config should prefer `baseUrl`.
|
|
|
|
## Auth rules
|
|
|
|
<AccordionGroup>
|
|
<Accordion title="Local and LAN hosts">
|
|
Local and LAN Ollama hosts do not need a real bearer token. OpenClaw uses the local `ollama-local` marker only for loopback, private-network, `.local`, and bare-hostname Ollama base URLs.
|
|
</Accordion>
|
|
<Accordion title="Remote and Ollama Cloud hosts">
|
|
Remote public hosts and Ollama Cloud (`https://ollama.com`) require a real credential through `OLLAMA_API_KEY`, an auth profile, or the provider's `apiKey`.
|
|
</Accordion>
|
|
<Accordion title="Custom provider ids">
|
|
Custom provider ids that set `api: "ollama"` follow the same rules. For example, an `ollama-remote` provider that points at a private LAN Ollama host can use `apiKey: "ollama-local"` and sub-agents will resolve that marker through the Ollama provider hook instead of treating it as a missing credential.
|
|
</Accordion>
|
|
<Accordion title="Memory embedding scope">
|
|
When Ollama is used for memory embeddings, bearer auth is scoped to the host where it was declared:
|
|
|
|
- A provider-level key is sent only to that provider's Ollama host.
|
|
- `agents.*.memorySearch.remote.apiKey` is sent only to its remote embedding host.
|
|
- A pure `OLLAMA_API_KEY` env value is treated as the Ollama Cloud convention, not sent to local or self-hosted hosts by default.
|
|
|
|
</Accordion>
|
|
</AccordionGroup>
|
|
|
|
## Getting started
|
|
|
|
Choose your preferred setup method and mode.
|
|
|
|
<Tabs>
|
|
<Tab title="Onboarding (recommended)">
|
|
**Best for:** fastest path to a working Ollama cloud or local setup.
|
|
|
|
<Steps>
|
|
<Step title="Run onboarding">
|
|
```bash
|
|
openclaw onboard
|
|
```
|
|
|
|
Select **Ollama** from the provider list.
|
|
</Step>
|
|
<Step title="Choose your mode">
|
|
- **Cloud + Local** — local Ollama host plus cloud models routed through that host
|
|
- **Cloud only** — hosted Ollama models via `https://ollama.com`
|
|
- **Local only** — local models only
|
|
</Step>
|
|
<Step title="Select a model">
|
|
`Cloud only` prompts for `OLLAMA_API_KEY` and suggests hosted cloud defaults. `Cloud + Local` and `Local only` ask for an Ollama base URL, discover available models, and auto-pull the selected local model if it is not available yet. `Cloud + Local` also checks whether that Ollama host is signed in for cloud access.
|
|
</Step>
|
|
<Step title="Verify the model is available">
|
|
```bash
|
|
openclaw models list --provider ollama
|
|
```
|
|
</Step>
|
|
</Steps>
|
|
|
|
### Non-interactive mode
|
|
|
|
```bash
|
|
openclaw onboard --non-interactive \
|
|
--auth-choice ollama \
|
|
--accept-risk
|
|
```
|
|
|
|
Optionally specify a custom base URL or model:
|
|
|
|
```bash
|
|
openclaw onboard --non-interactive \
|
|
--auth-choice ollama \
|
|
--custom-base-url "http://ollama-host:11434" \
|
|
--custom-model-id "qwen3.5:27b" \
|
|
--accept-risk
|
|
```
|
|
|
|
</Tab>
|
|
|
|
<Tab title="Manual setup">
|
|
**Best for:** full control over cloud or local setup.
|
|
|
|
<Steps>
|
|
<Step title="Choose cloud or local">
|
|
- **Cloud + Local**: install Ollama, sign in with `ollama signin`, and route cloud requests through that host
|
|
- **Cloud only**: use `https://ollama.com` with an `OLLAMA_API_KEY`
|
|
- **Local only**: install Ollama from [ollama.com/download](https://ollama.com/download)
|
|
</Step>
|
|
<Step title="Pull a local model (local only)">
|
|
```bash
|
|
ollama pull gemma4
|
|
# or
|
|
ollama pull gpt-oss:20b
|
|
# or
|
|
ollama pull llama3.3
|
|
```
|
|
</Step>
|
|
<Step title="Enable Ollama for OpenClaw">
|
|
For `Cloud only`, use your real `OLLAMA_API_KEY`. For host-backed setups, any placeholder value works:
|
|
|
|
```bash
|
|
# Cloud
|
|
export OLLAMA_API_KEY="your-ollama-api-key"
|
|
|
|
# Local-only
|
|
export OLLAMA_API_KEY="ollama-local"
|
|
|
|
# Or configure in your config file
|
|
openclaw config set models.providers.ollama.apiKey "OLLAMA_API_KEY"
|
|
```
|
|
</Step>
|
|
<Step title="Inspect and set your model">
|
|
```bash
|
|
openclaw models list
|
|
openclaw models set ollama/gemma4
|
|
```
|
|
|
|
Or set the default in config:
|
|
|
|
```json5
|
|
{
|
|
agents: {
|
|
defaults: {
|
|
model: { primary: "ollama/gemma4" },
|
|
},
|
|
},
|
|
}
|
|
```
|
|
</Step>
|
|
</Steps>
|
|
|
|
</Tab>
|
|
</Tabs>
|
|
|
|
## Cloud models
|
|
|
|
<Tabs>
|
|
<Tab title="Cloud + Local">
|
|
`Cloud + Local` uses a reachable Ollama host as the control point for both local and cloud models. This is Ollama's preferred hybrid flow.
|
|
|
|
Use **Cloud + Local** during setup. OpenClaw prompts for the Ollama base URL, discovers local models from that host, and checks whether the host is signed in for cloud access with `ollama signin`. When the host is signed in, OpenClaw also suggests hosted cloud defaults such as `kimi-k2.5:cloud`, `minimax-m2.7:cloud`, and `glm-5.1:cloud`.
|
|
|
|
If the host is not signed in yet, OpenClaw keeps the setup local-only until you run `ollama signin`.
|
|
|
|
</Tab>
|
|
|
|
<Tab title="Cloud only">
|
|
`Cloud only` runs against Ollama's hosted API at `https://ollama.com`.
|
|
|
|
Use **Cloud only** during setup. OpenClaw prompts for `OLLAMA_API_KEY`, sets `baseUrl: "https://ollama.com"`, and seeds the hosted cloud model list. This path does **not** require a local Ollama server or `ollama signin`.
|
|
|
|
The cloud model list shown during `openclaw onboard` is populated live from `https://ollama.com/api/tags`, capped at 500 entries, so the picker reflects the current hosted catalog rather than a static seed. If `ollama.com` is unreachable or returns no models at setup time, OpenClaw falls back to the previous hardcoded suggestions so onboarding still completes.
|
|
|
|
</Tab>
|
|
|
|
<Tab title="Local only">
|
|
In local-only mode, OpenClaw discovers models from the configured Ollama instance. This path is for local or self-hosted Ollama servers.
|
|
|
|
OpenClaw currently suggests `gemma4` as the local default.
|
|
|
|
</Tab>
|
|
</Tabs>
|
|
|
|
## Model discovery (implicit provider)
|
|
|
|
When you set `OLLAMA_API_KEY` (or an auth profile) and **do not** define `models.providers.ollama` or another custom remote provider with `api: "ollama"`, OpenClaw discovers models from the local Ollama instance at `http://127.0.0.1:11434`.
|
|
|
|
| Behavior | Detail |
|
|
| -------------------- | ------------------------------------------------------------------------------------------------------------------------------------------------------------------- |
|
|
| Catalog query | Queries `/api/tags` |
|
|
| Capability detection | Uses best-effort `/api/show` lookups to read `contextWindow`, expanded `num_ctx` Modelfile parameters, and capabilities including vision/tools |
|
|
| Vision models | Models with a `vision` capability reported by `/api/show` are marked as image-capable (`input: ["text", "image"]`), so OpenClaw auto-injects images into the prompt |
|
|
| Reasoning detection | Marks `reasoning` with a model-name heuristic (`r1`, `reasoning`, `think`) |
|
|
| Token limits | Sets `maxTokens` to the default Ollama max-token cap used by OpenClaw |
|
|
| Costs | Sets all costs to `0` |
|
|
|
|
This avoids manual model entries while keeping the catalog aligned with the local Ollama instance.
|
|
|
|
```bash
|
|
# See what models are available
|
|
ollama list
|
|
openclaw models list
|
|
```
|
|
|
|
To add a new model, simply pull it with Ollama:
|
|
|
|
```bash
|
|
ollama pull mistral
|
|
```
|
|
|
|
The new model will be automatically discovered and available to use.
|
|
|
|
<Note>
|
|
If you set `models.providers.ollama` explicitly, or configure a custom remote provider such as `models.providers.ollama-cloud` with `api: "ollama"`, auto-discovery is skipped and you must define models manually. Loopback custom providers such as `http://127.0.0.2:11434` are still treated as local. See the explicit config section below.
|
|
</Note>
|
|
|
|
## Vision and image description
|
|
|
|
The bundled Ollama plugin registers Ollama as an image-capable media-understanding provider. This lets OpenClaw route explicit image-description requests and configured image-model defaults through local or hosted Ollama vision models.
|
|
|
|
For local vision, pull a model that supports images:
|
|
|
|
```bash
|
|
ollama pull qwen2.5vl:7b
|
|
export OLLAMA_API_KEY="ollama-local"
|
|
```
|
|
|
|
Then verify with the infer CLI:
|
|
|
|
```bash
|
|
openclaw infer image describe \
|
|
--file ./photo.jpg \
|
|
--model ollama/qwen2.5vl:7b \
|
|
--json
|
|
```
|
|
|
|
`--model` must be a full `<provider/model>` ref. When it is set, `openclaw infer image describe` runs that model directly instead of skipping description because the model supports native vision.
|
|
|
|
To make Ollama the default image-understanding model for inbound media, configure `agents.defaults.imageModel`:
|
|
|
|
```json5
|
|
{
|
|
agents: {
|
|
defaults: {
|
|
imageModel: {
|
|
primary: "ollama/qwen2.5vl:7b",
|
|
},
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
If you define `models.providers.ollama.models` manually, mark vision models with image input support:
|
|
|
|
```json5
|
|
{
|
|
id: "qwen2.5vl:7b",
|
|
name: "qwen2.5vl:7b",
|
|
input: ["text", "image"],
|
|
contextWindow: 128000,
|
|
maxTokens: 8192,
|
|
}
|
|
```
|
|
|
|
OpenClaw rejects image-description requests for models that are not marked image-capable. With implicit discovery, OpenClaw reads this from Ollama when `/api/show` reports a vision capability.
|
|
|
|
## Configuration
|
|
|
|
<Tabs>
|
|
<Tab title="Basic (implicit discovery)">
|
|
The simplest local-only enablement path is via environment variable:
|
|
|
|
```bash
|
|
export OLLAMA_API_KEY="ollama-local"
|
|
```
|
|
|
|
<Tip>
|
|
If `OLLAMA_API_KEY` is set, you can omit `apiKey` in the provider entry and OpenClaw will fill it for availability checks.
|
|
</Tip>
|
|
|
|
</Tab>
|
|
|
|
<Tab title="Explicit (manual models)">
|
|
Use explicit config when you want hosted cloud setup, Ollama runs on another host/port, you want to force specific context windows or model lists, or you want fully manual model definitions.
|
|
|
|
```json5
|
|
{
|
|
models: {
|
|
providers: {
|
|
ollama: {
|
|
baseUrl: "https://ollama.com",
|
|
apiKey: "OLLAMA_API_KEY",
|
|
api: "ollama",
|
|
models: [
|
|
{
|
|
id: "kimi-k2.5:cloud",
|
|
name: "kimi-k2.5:cloud",
|
|
reasoning: false,
|
|
input: ["text", "image"],
|
|
cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 },
|
|
contextWindow: 128000,
|
|
maxTokens: 8192
|
|
}
|
|
]
|
|
}
|
|
}
|
|
}
|
|
}
|
|
```
|
|
|
|
</Tab>
|
|
|
|
<Tab title="Custom base URL">
|
|
If Ollama is running on a different host or port (explicit config disables auto-discovery, so define models manually):
|
|
|
|
```json5
|
|
{
|
|
models: {
|
|
providers: {
|
|
ollama: {
|
|
apiKey: "ollama-local",
|
|
baseUrl: "http://ollama-host:11434", // No /v1 - use native Ollama API URL
|
|
api: "ollama", // Set explicitly to guarantee native tool-calling behavior
|
|
timeoutSeconds: 300, // Optional: give cold local models longer to connect and stream
|
|
models: [
|
|
{
|
|
id: "qwen3:32b",
|
|
name: "qwen3:32b",
|
|
params: {
|
|
keep_alive: "15m", // Optional: keep the model loaded between turns
|
|
},
|
|
},
|
|
],
|
|
},
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
<Warning>
|
|
Do not add `/v1` to the URL. The `/v1` path uses OpenAI-compatible mode, where tool calling is not reliable. Use the base Ollama URL without a path suffix.
|
|
</Warning>
|
|
|
|
</Tab>
|
|
</Tabs>
|
|
|
|
## Common recipes
|
|
|
|
Use these as starting points and replace model IDs with the exact names from `ollama list` or `openclaw models list --provider ollama`.
|
|
|
|
<AccordionGroup>
|
|
<Accordion title="Local model with auto-discovery">
|
|
Use this when Ollama runs on the same machine as the Gateway and you want OpenClaw to discover the installed models automatically.
|
|
|
|
```bash
|
|
ollama serve
|
|
ollama pull gemma4
|
|
export OLLAMA_API_KEY="ollama-local"
|
|
openclaw models list --provider ollama
|
|
openclaw models set ollama/gemma4
|
|
```
|
|
|
|
This path keeps config minimal. Do not add a `models.providers.ollama` block unless you want to define models manually.
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="LAN Ollama host with manual models">
|
|
Use native Ollama URLs for LAN hosts. Do not add `/v1`.
|
|
|
|
```json5
|
|
{
|
|
models: {
|
|
providers: {
|
|
ollama: {
|
|
baseUrl: "http://gpu-box.local:11434",
|
|
apiKey: "ollama-local",
|
|
api: "ollama",
|
|
timeoutSeconds: 300,
|
|
contextWindow: 32768,
|
|
maxTokens: 8192,
|
|
models: [
|
|
{
|
|
id: "qwen3.5:9b",
|
|
name: "qwen3.5:9b",
|
|
reasoning: true,
|
|
input: ["text"],
|
|
params: {
|
|
num_ctx: 32768,
|
|
keep_alive: "15m",
|
|
},
|
|
},
|
|
],
|
|
},
|
|
},
|
|
},
|
|
agents: {
|
|
defaults: {
|
|
model: { primary: "ollama/qwen3.5:9b" },
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
`contextWindow` is the OpenClaw-side context budget. `params.num_ctx` is sent to Ollama for the request. Keep them aligned when your hardware cannot run the model's full advertised context.
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="Ollama Cloud only">
|
|
Use this when you do not run a local daemon and want hosted Ollama models directly.
|
|
|
|
```bash
|
|
export OLLAMA_API_KEY="your-ollama-api-key"
|
|
```
|
|
|
|
```json5
|
|
{
|
|
models: {
|
|
providers: {
|
|
ollama: {
|
|
baseUrl: "https://ollama.com",
|
|
apiKey: "OLLAMA_API_KEY",
|
|
api: "ollama",
|
|
models: [
|
|
{
|
|
id: "kimi-k2.5:cloud",
|
|
name: "kimi-k2.5:cloud",
|
|
reasoning: false,
|
|
input: ["text", "image"],
|
|
contextWindow: 128000,
|
|
maxTokens: 8192,
|
|
},
|
|
],
|
|
},
|
|
},
|
|
},
|
|
agents: {
|
|
defaults: {
|
|
model: { primary: "ollama/kimi-k2.5:cloud" },
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="Cloud plus local through a signed-in daemon">
|
|
Use this when a local or LAN Ollama daemon is signed in with `ollama signin` and should serve both local models and `:cloud` models.
|
|
|
|
```bash
|
|
ollama signin
|
|
ollama pull gemma4
|
|
```
|
|
|
|
```json5
|
|
{
|
|
models: {
|
|
providers: {
|
|
ollama: {
|
|
baseUrl: "http://127.0.0.1:11434",
|
|
apiKey: "ollama-local",
|
|
api: "ollama",
|
|
timeoutSeconds: 300,
|
|
models: [
|
|
{ id: "gemma4", name: "gemma4", input: ["text"] },
|
|
{ id: "kimi-k2.5:cloud", name: "kimi-k2.5:cloud", input: ["text", "image"] },
|
|
],
|
|
},
|
|
},
|
|
},
|
|
agents: {
|
|
defaults: {
|
|
model: {
|
|
primary: "ollama/gemma4",
|
|
fallbacks: ["ollama/kimi-k2.5:cloud"],
|
|
},
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="Multiple Ollama hosts">
|
|
Use custom provider IDs when you have more than one Ollama server. Each provider gets its own host, models, auth, timeout, and model refs.
|
|
|
|
```json5
|
|
{
|
|
models: {
|
|
providers: {
|
|
"ollama-fast": {
|
|
baseUrl: "http://mini.local:11434",
|
|
apiKey: "ollama-local",
|
|
api: "ollama",
|
|
contextWindow: 32768,
|
|
models: [{ id: "gemma4", name: "gemma4", input: ["text"] }],
|
|
},
|
|
"ollama-large": {
|
|
baseUrl: "http://gpu-box.local:11434",
|
|
apiKey: "ollama-local",
|
|
api: "ollama",
|
|
timeoutSeconds: 420,
|
|
contextWindow: 131072,
|
|
maxTokens: 16384,
|
|
models: [{ id: "qwen3.5:27b", name: "qwen3.5:27b", input: ["text"] }],
|
|
},
|
|
},
|
|
},
|
|
agents: {
|
|
defaults: {
|
|
model: {
|
|
primary: "ollama-fast/gemma4",
|
|
fallbacks: ["ollama-large/qwen3.5:27b"],
|
|
},
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
When OpenClaw sends the request, the active provider prefix is stripped so `ollama-large/qwen3.5:27b` reaches Ollama as `qwen3.5:27b`.
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="Lean local model profile">
|
|
Some local models can answer simple prompts but struggle with the full agent tool surface. Start by limiting tools and context before changing global runtime settings.
|
|
|
|
```json5
|
|
{
|
|
agents: {
|
|
defaults: {
|
|
experimental: {
|
|
localModelLean: true,
|
|
},
|
|
model: { primary: "ollama/gemma4" },
|
|
},
|
|
},
|
|
models: {
|
|
providers: {
|
|
ollama: {
|
|
baseUrl: "http://127.0.0.1:11434",
|
|
apiKey: "ollama-local",
|
|
api: "ollama",
|
|
contextWindow: 32768,
|
|
models: [
|
|
{
|
|
id: "gemma4",
|
|
name: "gemma4",
|
|
input: ["text"],
|
|
params: { num_ctx: 32768 },
|
|
compat: { supportsTools: false },
|
|
},
|
|
],
|
|
},
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
Use `compat.supportsTools: false` only when the model or server reliably fails on tool schemas. It trades agent capability for stability.
|
|
|
|
</Accordion>
|
|
</AccordionGroup>
|
|
|
|
### Model selection
|
|
|
|
Once configured, all your Ollama models are available:
|
|
|
|
```json5
|
|
{
|
|
agents: {
|
|
defaults: {
|
|
model: {
|
|
primary: "ollama/gpt-oss:20b",
|
|
fallbacks: ["ollama/llama3.3", "ollama/qwen2.5-coder:32b"],
|
|
},
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
Custom Ollama provider ids are also supported. When a model ref uses the active
|
|
provider prefix, such as `ollama-spark/qwen3:32b`, OpenClaw strips only that
|
|
prefix before calling Ollama so the server receives `qwen3:32b`.
|
|
|
|
For slow local models, prefer provider-scoped request tuning before raising the
|
|
whole agent runtime timeout:
|
|
|
|
```json5
|
|
{
|
|
models: {
|
|
providers: {
|
|
ollama: {
|
|
timeoutSeconds: 300,
|
|
models: [
|
|
{
|
|
id: "gemma4:26b",
|
|
name: "gemma4:26b",
|
|
params: { keep_alive: "15m" },
|
|
},
|
|
],
|
|
},
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
`timeoutSeconds` applies to the model HTTP request, including connection setup,
|
|
headers, body streaming, and the total guarded-fetch abort. `params.keep_alive`
|
|
is forwarded to Ollama as top-level `keep_alive` on native `/api/chat` requests;
|
|
set it per model when first-turn load time is the bottleneck.
|
|
|
|
### Quick verification
|
|
|
|
```bash
|
|
# Ollama daemon visible to this machine
|
|
curl http://127.0.0.1:11434/api/tags
|
|
|
|
# OpenClaw catalog and selected model
|
|
openclaw models list --provider ollama
|
|
openclaw models status
|
|
|
|
# Direct model smoke
|
|
openclaw infer model run \
|
|
--model ollama/gemma4 \
|
|
--prompt "Reply with exactly: ok"
|
|
```
|
|
|
|
For remote hosts, replace `127.0.0.1` with the host used in `baseUrl`. If `curl` works but OpenClaw does not, check whether the Gateway runs on a different machine, container, or service account.
|
|
|
|
## Ollama Web Search
|
|
|
|
OpenClaw supports **Ollama Web Search** as a bundled `web_search` provider.
|
|
|
|
| Property | Detail |
|
|
| ----------- | -------------------------------------------------------------------------------------------------------------------------------------------------------------------- |
|
|
| Host | Uses your configured Ollama host (`models.providers.ollama.baseUrl` when set, otherwise `http://127.0.0.1:11434`); `https://ollama.com` uses the hosted API directly |
|
|
| Auth | Key-free for signed-in local Ollama hosts; `OLLAMA_API_KEY` or configured provider auth for direct `https://ollama.com` search or auth-protected hosts |
|
|
| Requirement | Local/self-hosted hosts must be running and signed in with `ollama signin`; direct hosted search requires `baseUrl: "https://ollama.com"` plus a real Ollama API key |
|
|
|
|
Choose **Ollama Web Search** during `openclaw onboard` or `openclaw configure --section web`, or set:
|
|
|
|
```json5
|
|
{
|
|
tools: {
|
|
web: {
|
|
search: {
|
|
provider: "ollama",
|
|
},
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
For direct hosted search through Ollama Cloud:
|
|
|
|
```json5
|
|
{
|
|
models: {
|
|
providers: {
|
|
ollama: {
|
|
baseUrl: "https://ollama.com",
|
|
apiKey: "OLLAMA_API_KEY",
|
|
api: "ollama",
|
|
models: [{ id: "kimi-k2.5:cloud", name: "kimi-k2.5:cloud", input: ["text"] }],
|
|
},
|
|
},
|
|
},
|
|
tools: {
|
|
web: {
|
|
search: { provider: "ollama" },
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
For a signed-in local daemon, OpenClaw uses the daemon's `/api/experimental/web_search` proxy. For `https://ollama.com`, it calls the hosted `/api/web_search` endpoint directly.
|
|
|
|
<Note>
|
|
For the full setup and behavior details, see [Ollama Web Search](/tools/ollama-search).
|
|
</Note>
|
|
|
|
## Advanced configuration
|
|
|
|
<AccordionGroup>
|
|
<Accordion title="Legacy OpenAI-compatible mode">
|
|
<Warning>
|
|
**Tool calling is not reliable in OpenAI-compatible mode.** Use this mode only if you need OpenAI format for a proxy and do not depend on native tool calling behavior.
|
|
</Warning>
|
|
|
|
If you need to use the OpenAI-compatible endpoint instead (for example, behind a proxy that only supports OpenAI format), set `api: "openai-completions"` explicitly:
|
|
|
|
```json5
|
|
{
|
|
models: {
|
|
providers: {
|
|
ollama: {
|
|
baseUrl: "http://ollama-host:11434/v1",
|
|
api: "openai-completions",
|
|
injectNumCtxForOpenAICompat: true, // default: true
|
|
apiKey: "ollama-local",
|
|
models: [...]
|
|
}
|
|
}
|
|
}
|
|
}
|
|
```
|
|
|
|
This mode may not support streaming and tool calling simultaneously. You may need to disable streaming with `params: { streaming: false }` in model config.
|
|
|
|
When `api: "openai-completions"` is used with Ollama, OpenClaw injects `options.num_ctx` by default so Ollama does not silently fall back to a 4096 context window. If your proxy/upstream rejects unknown `options` fields, disable this behavior:
|
|
|
|
```json5
|
|
{
|
|
models: {
|
|
providers: {
|
|
ollama: {
|
|
baseUrl: "http://ollama-host:11434/v1",
|
|
api: "openai-completions",
|
|
injectNumCtxForOpenAICompat: false,
|
|
apiKey: "ollama-local",
|
|
models: [...]
|
|
}
|
|
}
|
|
}
|
|
}
|
|
```
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="Context windows">
|
|
For auto-discovered models, OpenClaw uses the context window reported by Ollama when available, including larger `PARAMETER num_ctx` values from custom Modelfiles. Otherwise it falls back to the default Ollama context window used by OpenClaw.
|
|
|
|
You can set provider-level `contextWindow`, `contextTokens`, and `maxTokens` defaults for every model under that Ollama provider, then override them per model when needed. To cap Ollama's per-request runtime context without rebuilding a Modelfile, set `params.num_ctx`; OpenClaw sends it as `options.num_ctx` for both native Ollama and the OpenAI-compatible Ollama adapter. Invalid, zero, negative, and non-finite values are ignored and fall back to `contextWindow`.
|
|
|
|
Native Ollama model entries also accept the common Ollama runtime options under `params`, including `temperature`, `top_p`, `top_k`, `min_p`, `num_predict`, `stop`, `repeat_penalty`, `num_batch`, `num_thread`, and `use_mmap`. OpenClaw forwards only Ollama request keys, so OpenClaw runtime params such as `streaming` are not leaked to Ollama. Use `params.think` or `params.thinking` to send top-level Ollama `think`; `false` disables API-level thinking for Qwen-style thinking models.
|
|
|
|
```json5
|
|
{
|
|
models: {
|
|
providers: {
|
|
ollama: {
|
|
contextWindow: 32768,
|
|
models: [
|
|
{
|
|
id: "llama3.3",
|
|
contextWindow: 131072,
|
|
maxTokens: 65536,
|
|
params: {
|
|
num_ctx: 32768,
|
|
temperature: 0.7,
|
|
top_p: 0.9,
|
|
thinking: false,
|
|
},
|
|
}
|
|
]
|
|
}
|
|
}
|
|
}
|
|
}
|
|
```
|
|
|
|
Per-model `agents.defaults.models["ollama/<model>"].params.num_ctx` works too. If both are configured, the explicit provider model entry wins over the agent default.
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="Thinking control">
|
|
For native Ollama models, OpenClaw forwards thinking control as Ollama expects it: top-level `think`, not `options.think`.
|
|
|
|
```bash
|
|
openclaw agent --model ollama/gemma4 --thinking off
|
|
openclaw agent --model ollama/gemma4 --thinking low
|
|
```
|
|
|
|
You can also set a model default:
|
|
|
|
```json5
|
|
{
|
|
agents: {
|
|
defaults: {
|
|
models: {
|
|
"ollama/gemma4": {
|
|
thinking: "low",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
Per-model `params.think` or `params.thinking` can disable or force Ollama API thinking for a specific configured model. Runtime commands such as `/think off` still apply to the active run.
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="Reasoning models">
|
|
OpenClaw treats models with names such as `deepseek-r1`, `reasoning`, or `think` as reasoning-capable by default.
|
|
|
|
```bash
|
|
ollama pull deepseek-r1:32b
|
|
```
|
|
|
|
No additional configuration is needed. OpenClaw marks them automatically.
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="Model costs">
|
|
Ollama is free and runs locally, so all model costs are set to $0. This applies to both auto-discovered and manually defined models.
|
|
</Accordion>
|
|
|
|
<Accordion title="Memory embeddings">
|
|
The bundled Ollama plugin registers a memory embedding provider for
|
|
[memory search](/concepts/memory). It uses the configured Ollama base URL
|
|
and API key, calls Ollama's current `/api/embed` endpoint, and batches
|
|
multiple memory chunks into one `input` request when possible.
|
|
|
|
| Property | Value |
|
|
| ------------- | ------------------- |
|
|
| Default model | `nomic-embed-text` |
|
|
| Auto-pull | Yes — the embedding model is pulled automatically if not present locally |
|
|
|
|
To select Ollama as the memory search embedding provider:
|
|
|
|
```json5
|
|
{
|
|
agents: {
|
|
defaults: {
|
|
memorySearch: { provider: "ollama" },
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
For a remote embedding host, keep auth scoped to that host:
|
|
|
|
```json5
|
|
{
|
|
agents: {
|
|
defaults: {
|
|
memorySearch: {
|
|
provider: "ollama",
|
|
remote: {
|
|
baseUrl: "http://gpu-box.local:11434",
|
|
model: "nomic-embed-text",
|
|
apiKey: "ollama-local",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="Streaming configuration">
|
|
OpenClaw's Ollama integration uses the **native Ollama API** (`/api/chat`) by default, which fully supports streaming and tool calling simultaneously. No special configuration is needed.
|
|
|
|
For native `/api/chat` requests, OpenClaw also forwards thinking control directly to Ollama: `/think off` and `openclaw agent --thinking off` send top-level `think: false`, while `/think low|medium|high` send the matching top-level `think` effort string. `/think max` maps to Ollama's highest native effort, `think: "high"`.
|
|
|
|
<Tip>
|
|
If you need to use the OpenAI-compatible endpoint, see the "Legacy OpenAI-compatible mode" section above. Streaming and tool calling may not work simultaneously in that mode.
|
|
</Tip>
|
|
|
|
</Accordion>
|
|
</AccordionGroup>
|
|
|
|
## Troubleshooting
|
|
|
|
<AccordionGroup>
|
|
<Accordion title="Ollama not detected">
|
|
Make sure Ollama is running and that you set `OLLAMA_API_KEY` (or an auth profile), and that you did **not** define an explicit `models.providers.ollama` entry:
|
|
|
|
```bash
|
|
ollama serve
|
|
```
|
|
|
|
Verify that the API is accessible:
|
|
|
|
```bash
|
|
curl http://localhost:11434/api/tags
|
|
```
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="No models available">
|
|
If your model is not listed, either pull the model locally or define it explicitly in `models.providers.ollama`.
|
|
|
|
```bash
|
|
ollama list # See what's installed
|
|
ollama pull gemma4
|
|
ollama pull gpt-oss:20b
|
|
ollama pull llama3.3 # Or another model
|
|
```
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="Connection refused">
|
|
Check that Ollama is running on the correct port:
|
|
|
|
```bash
|
|
# Check if Ollama is running
|
|
ps aux | grep ollama
|
|
|
|
# Or restart Ollama
|
|
ollama serve
|
|
```
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="Remote host works with curl but not OpenClaw">
|
|
Verify from the same machine and runtime that runs the Gateway:
|
|
|
|
```bash
|
|
openclaw gateway status --deep
|
|
curl http://ollama-host:11434/api/tags
|
|
```
|
|
|
|
Common causes:
|
|
|
|
- `baseUrl` points at `localhost`, but the Gateway runs in Docker or on another host.
|
|
- The URL uses `/v1`, which selects OpenAI-compatible behavior instead of native Ollama.
|
|
- The remote host needs firewall or LAN binding changes on the Ollama side.
|
|
- The model is present on your laptop's daemon but not on the remote daemon.
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="Model outputs tool JSON as text">
|
|
This usually means the provider is using OpenAI-compatible mode or the model cannot handle tool schemas.
|
|
|
|
Prefer native Ollama mode:
|
|
|
|
```json5
|
|
{
|
|
models: {
|
|
providers: {
|
|
ollama: {
|
|
baseUrl: "http://ollama-host:11434",
|
|
api: "ollama",
|
|
},
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
If a small local model still fails on tool schemas, set `compat.supportsTools: false` on that model entry and retest.
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="Cold local model times out">
|
|
Large local models can need a long first load before streaming begins. Keep the timeout scoped to the Ollama provider, and optionally ask Ollama to keep the model loaded between turns:
|
|
|
|
```json5
|
|
{
|
|
models: {
|
|
providers: {
|
|
ollama: {
|
|
timeoutSeconds: 300,
|
|
models: [
|
|
{
|
|
id: "gemma4:26b",
|
|
name: "gemma4:26b",
|
|
params: { keep_alive: "15m" },
|
|
},
|
|
],
|
|
},
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
If the host itself is slow to accept connections, `timeoutSeconds` also extends the guarded Undici connect timeout for this provider.
|
|
|
|
</Accordion>
|
|
|
|
<Accordion title="Large-context model is too slow or runs out of memory">
|
|
Many Ollama models advertise contexts that are larger than your hardware can run comfortably. Cap both OpenClaw's budget and Ollama's request context:
|
|
|
|
```json5
|
|
{
|
|
models: {
|
|
providers: {
|
|
ollama: {
|
|
contextWindow: 32768,
|
|
maxTokens: 8192,
|
|
models: [
|
|
{
|
|
id: "qwen3.5:9b",
|
|
name: "qwen3.5:9b",
|
|
params: { num_ctx: 32768 },
|
|
},
|
|
],
|
|
},
|
|
},
|
|
},
|
|
}
|
|
```
|
|
|
|
Lower `contextWindow` first if the prompt ingestion phase is slow. Lower `maxTokens` if generation runs too long.
|
|
|
|
</Accordion>
|
|
</AccordionGroup>
|
|
|
|
<Note>
|
|
More help: [Troubleshooting](/help/troubleshooting) and [FAQ](/help/faq).
|
|
</Note>
|
|
|
|
## Related
|
|
|
|
<CardGroup cols={2}>
|
|
<Card title="Model providers" href="/concepts/model-providers" icon="layers">
|
|
Overview of all providers, model refs, and failover behavior.
|
|
</Card>
|
|
<Card title="Model selection" href="/concepts/models" icon="brain">
|
|
How to choose and configure models.
|
|
</Card>
|
|
<Card title="Ollama Web Search" href="/tools/ollama-search" icon="magnifying-glass">
|
|
Full setup and behavior details for Ollama-powered web search.
|
|
</Card>
|
|
<Card title="Configuration" href="/gateway/configuration" icon="gear">
|
|
Full config reference.
|
|
</Card>
|
|
</CardGroup>
|