mirror of
https://github.com/openclaw/openclaw.git
synced 2026-05-06 09:10:45 +00:00
fix: clarify memory embedding concurrency help
This commit is contained in:
@@ -1,4 +1,4 @@
|
||||
0f57fb6d20b9d300c4325b227e49f17f04349b0f3c27dd218397fe7a3b5001dc config-baseline.json
|
||||
9d1815981dc3f89d1dfdc72f0a4723d4fd5efca8e5b8a1a1cbf6a053c50c937d config-baseline.core.json
|
||||
87b6a979987a6f6a47c16e1cbfc924d3b7cc9cd68139e9e1460e6e263c877052 config-baseline.json
|
||||
7937564a6c8020b765b857b52b522beaa24d970f5743833716cd019b7147de10 config-baseline.core.json
|
||||
c4f07c228d4f07e7afafa5b600b4a80f5b26aaed7267c7287a64d04a527be8e8 config-baseline.channel.json
|
||||
6938050627f0d120109d2045b4300aa8b508b35132542db434033ed0fe3e2b3a config-baseline.plugin.json
|
||||
|
||||
@@ -4374,7 +4374,7 @@ export const GENERATED_BASE_CONFIG_SCHEMA: BaseConfigSchemaResponse = {
|
||||
maximum: 9007199254740991,
|
||||
title: "Remote Non-Batch Embedding Concurrency",
|
||||
description:
|
||||
"Limits concurrent inline embedding requests during non-batch memory indexing. Lower this for local or small self-hosted providers such as Ollama; batch embedding concurrency is configured separately under remote.batch.",
|
||||
"Controls concurrent inline embedding requests during non-batch memory indexing. Use a low value for local or small self-hosted providers such as Ollama; batch embedding concurrency is configured separately under remote.batch.",
|
||||
},
|
||||
batch: {
|
||||
type: "object",
|
||||
@@ -26096,7 +26096,7 @@ export const GENERATED_BASE_CONFIG_SCHEMA: BaseConfigSchemaResponse = {
|
||||
},
|
||||
"agents.defaults.memorySearch.remote.nonBatchConcurrency": {
|
||||
label: "Remote Non-Batch Embedding Concurrency",
|
||||
help: "Limits concurrent inline embedding requests during non-batch memory indexing. Lower this for local or small self-hosted providers such as Ollama; batch embedding concurrency is configured separately under remote.batch.",
|
||||
help: "Controls concurrent inline embedding requests during non-batch memory indexing. Use a low value for local or small self-hosted providers such as Ollama; batch embedding concurrency is configured separately under remote.batch.",
|
||||
tags: ["performance"],
|
||||
},
|
||||
"agents.defaults.memorySearch.remote.batch.enabled": {
|
||||
|
||||
@@ -1015,7 +1015,7 @@ export const FIELD_HELP: Record<string, string> = {
|
||||
"agents.defaults.memorySearch.remote.headers":
|
||||
"Adds custom HTTP headers to remote embedding requests, merged with provider defaults. Use this for proxy auth and tenant routing headers, and keep values minimal to avoid leaking sensitive metadata.",
|
||||
"agents.defaults.memorySearch.remote.nonBatchConcurrency":
|
||||
"Limits concurrent inline embedding requests during non-batch memory indexing. Lower this for local or small self-hosted providers such as Ollama; batch embedding concurrency is configured separately under remote.batch.",
|
||||
"Controls concurrent inline embedding requests during non-batch memory indexing. Use a low value for local or small self-hosted providers such as Ollama; batch embedding concurrency is configured separately under remote.batch.",
|
||||
"agents.defaults.memorySearch.remote.batch.enabled":
|
||||
"Enables provider batch APIs for embedding jobs when supported (OpenAI/Gemini), improving throughput on larger index runs. Keep this enabled unless debugging provider batch failures or running very small workloads.",
|
||||
"agents.defaults.memorySearch.remote.batch.wait":
|
||||
|
||||
Reference in New Issue
Block a user