docs: prune recent additions for readability

This commit is contained in:
Vincent Koc
2026-04-23 00:22:58 -07:00
parent 252e4dde39
commit daaedf37b7
18 changed files with 38 additions and 103 deletions

View File

@@ -104,9 +104,11 @@ Interactive setup can prompt for an optional preferred load context length and a
### Streaming usage compatibility
OpenClaw marks LM Studio as streaming-usage compatible, so token accounting no longer degrades to unknown or stale totals on streamed completions. OpenClaw also recovers token counts from llama.cpp-style `timings.prompt_n` / `timings.predicted_n` metadata when LM Studio does not emit an OpenAI-shaped `usage` object.
LM Studio is streaming-usage compatible. When it does not emit an OpenAI-shaped
`usage` object, OpenClaw recovers token counts from llama.cpp-style
`timings.prompt_n` / `timings.predicted_n` metadata instead.
Other OpenAI-compatible local backends covered by the same behavior:
Same behavior applies to these OpenAI-compatible local backends:
- vLLM
- SGLang