mirror of
https://github.com/openclaw/openclaw.git
synced 2026-05-07 10:20:42 +00:00
* Gateway: fix chat completions usage compatibility * Gateway: clarify usage-gated stream wait * Gateway: preserve aggregate usage totals * Agents: clamp usage components before total * fix(gateway): bound usage stream finalization * fix: add OpenAI compat usage changelog (#62986) (thanks @Lellansin) * fix(agents): emit lifecycle terminal events after flush --------- Co-authored-by: Ayaan Zaidi <hi@obviy.us>
311 lines
7.7 KiB
TypeScript
311 lines
7.7 KiB
TypeScript
import { describe, expect, it } from "vitest";
|
|
import {
|
|
normalizeUsage,
|
|
hasNonzeroUsage,
|
|
derivePromptTokens,
|
|
deriveSessionTotalTokens,
|
|
toOpenAiChatCompletionsUsage,
|
|
} from "./usage.js";
|
|
|
|
describe("normalizeUsage", () => {
|
|
it("normalizes cache fields from provider response", () => {
|
|
const usage = normalizeUsage({
|
|
input: 1000,
|
|
output: 500,
|
|
cacheRead: 2000,
|
|
cacheWrite: 300,
|
|
});
|
|
expect(usage).toEqual({
|
|
input: 1000,
|
|
output: 500,
|
|
cacheRead: 2000,
|
|
cacheWrite: 300,
|
|
total: undefined,
|
|
});
|
|
});
|
|
|
|
it("normalizes cache fields from alternate naming", () => {
|
|
const usage = normalizeUsage({
|
|
input_tokens: 1000,
|
|
output_tokens: 500,
|
|
cache_read_input_tokens: 2000,
|
|
cache_creation_input_tokens: 300,
|
|
});
|
|
expect(usage).toEqual({
|
|
input: 1000,
|
|
output: 500,
|
|
cacheRead: 2000,
|
|
cacheWrite: 300,
|
|
total: undefined,
|
|
});
|
|
});
|
|
|
|
it("handles cache_read and cache_write naming variants", () => {
|
|
const usage = normalizeUsage({
|
|
input: 1000,
|
|
cache_read: 1500,
|
|
cache_write: 200,
|
|
});
|
|
expect(usage).toEqual({
|
|
input: 1000,
|
|
output: undefined,
|
|
cacheRead: 1500,
|
|
cacheWrite: 200,
|
|
total: undefined,
|
|
});
|
|
});
|
|
|
|
it("handles Moonshot/Kimi cached_tokens field", () => {
|
|
// Moonshot v1 returns cached_tokens instead of cache_read_input_tokens
|
|
const usage = normalizeUsage({
|
|
prompt_tokens: 30,
|
|
completion_tokens: 9,
|
|
total_tokens: 39,
|
|
cached_tokens: 19,
|
|
});
|
|
expect(usage).toEqual({
|
|
input: 11,
|
|
output: 9,
|
|
cacheRead: 19,
|
|
cacheWrite: undefined,
|
|
total: 39,
|
|
});
|
|
});
|
|
|
|
it("handles Kimi K2 prompt_tokens_details.cached_tokens field", () => {
|
|
// Kimi K2 uses automatic prefix caching and returns cached_tokens in prompt_tokens_details
|
|
const usage = normalizeUsage({
|
|
prompt_tokens: 1113,
|
|
completion_tokens: 5,
|
|
total_tokens: 1118,
|
|
prompt_tokens_details: { cached_tokens: 1024 },
|
|
});
|
|
expect(usage).toEqual({
|
|
input: 89,
|
|
output: 5,
|
|
cacheRead: 1024,
|
|
cacheWrite: undefined,
|
|
total: 1118,
|
|
});
|
|
});
|
|
|
|
it("handles OpenAI Responses input_tokens_details.cached_tokens field", () => {
|
|
const usage = normalizeUsage({
|
|
input_tokens: 120,
|
|
output_tokens: 30,
|
|
total_tokens: 250,
|
|
input_tokens_details: { cached_tokens: 100 },
|
|
});
|
|
expect(usage).toEqual({
|
|
input: 20,
|
|
output: 30,
|
|
cacheRead: 100,
|
|
cacheWrite: undefined,
|
|
total: 250,
|
|
});
|
|
});
|
|
|
|
it("clamps negative input to zero (pre-subtracted cached_tokens > prompt_tokens)", () => {
|
|
// pi-ai OpenAI-format providers subtract cached_tokens from prompt_tokens
|
|
// upstream. When cached_tokens exceeds prompt_tokens the result is negative.
|
|
const usage = normalizeUsage({
|
|
input: -4900,
|
|
output: 200,
|
|
cacheRead: 5000,
|
|
});
|
|
expect(usage).toEqual({
|
|
input: 0,
|
|
output: 200,
|
|
cacheRead: 5000,
|
|
cacheWrite: undefined,
|
|
total: undefined,
|
|
});
|
|
});
|
|
|
|
it("clamps negative prompt_tokens alias to zero", () => {
|
|
const usage = normalizeUsage({
|
|
prompt_tokens: -12,
|
|
completion_tokens: 4,
|
|
});
|
|
expect(usage).toEqual({
|
|
input: 0,
|
|
output: 4,
|
|
cacheRead: undefined,
|
|
cacheWrite: undefined,
|
|
total: undefined,
|
|
});
|
|
});
|
|
|
|
it("returns undefined when no valid fields are provided", () => {
|
|
const usage = normalizeUsage(null);
|
|
expect(usage).toBeUndefined();
|
|
});
|
|
|
|
it("handles undefined input", () => {
|
|
const usage = normalizeUsage(undefined);
|
|
expect(usage).toBeUndefined();
|
|
});
|
|
});
|
|
|
|
describe("toOpenAiChatCompletionsUsage", () => {
|
|
it("uses max(component sum, aggregate total) when breakdown is partial", () => {
|
|
const usage = normalizeUsage({ output_tokens: 20, total_tokens: 100 });
|
|
expect(toOpenAiChatCompletionsUsage(usage)).toEqual({
|
|
prompt_tokens: 0,
|
|
completion_tokens: 20,
|
|
total_tokens: 100,
|
|
});
|
|
});
|
|
|
|
it("uses component sum when it exceeds aggregate total", () => {
|
|
expect(
|
|
toOpenAiChatCompletionsUsage({
|
|
input: 30,
|
|
output: 40,
|
|
total: 50,
|
|
}),
|
|
).toEqual({
|
|
prompt_tokens: 30,
|
|
completion_tokens: 40,
|
|
total_tokens: 70,
|
|
});
|
|
});
|
|
|
|
it("uses aggregate total when only total is present", () => {
|
|
const usage = normalizeUsage({ total_tokens: 42 });
|
|
expect(toOpenAiChatCompletionsUsage(usage)).toEqual({
|
|
prompt_tokens: 0,
|
|
completion_tokens: 0,
|
|
total_tokens: 42,
|
|
});
|
|
});
|
|
|
|
it("returns zeros for undefined usage", () => {
|
|
expect(toOpenAiChatCompletionsUsage(undefined)).toEqual({
|
|
prompt_tokens: 0,
|
|
completion_tokens: 0,
|
|
total_tokens: 0,
|
|
});
|
|
});
|
|
|
|
it("raises total_tokens with aggregate when cache write is excluded from prompt sum", () => {
|
|
expect(
|
|
toOpenAiChatCompletionsUsage({
|
|
input: 10,
|
|
output: 5,
|
|
cacheWrite: 100,
|
|
total: 200,
|
|
}),
|
|
).toEqual({
|
|
prompt_tokens: 10,
|
|
completion_tokens: 5,
|
|
total_tokens: 200,
|
|
});
|
|
});
|
|
|
|
it("clamps negative completion before deriving total_tokens", () => {
|
|
expect(
|
|
toOpenAiChatCompletionsUsage({
|
|
input: 3,
|
|
output: -5,
|
|
}),
|
|
).toEqual({
|
|
prompt_tokens: 3,
|
|
completion_tokens: 0,
|
|
total_tokens: 3,
|
|
});
|
|
});
|
|
|
|
it("preserves aggregate total when components are partially negative", () => {
|
|
expect(
|
|
toOpenAiChatCompletionsUsage({
|
|
input: 3,
|
|
output: -5,
|
|
total: 7,
|
|
}),
|
|
).toEqual({
|
|
prompt_tokens: 3,
|
|
completion_tokens: 0,
|
|
total_tokens: 7,
|
|
});
|
|
});
|
|
});
|
|
|
|
describe("hasNonzeroUsage", () => {
|
|
it("returns true when cache read is nonzero", () => {
|
|
const usage = { cacheRead: 100 };
|
|
expect(hasNonzeroUsage(usage)).toBe(true);
|
|
});
|
|
|
|
it("returns true when cache write is nonzero", () => {
|
|
const usage = { cacheWrite: 50 };
|
|
expect(hasNonzeroUsage(usage)).toBe(true);
|
|
});
|
|
|
|
it("returns true when both cache fields are nonzero", () => {
|
|
const usage = { cacheRead: 100, cacheWrite: 50 };
|
|
expect(hasNonzeroUsage(usage)).toBe(true);
|
|
});
|
|
|
|
it("returns false when cache fields are zero", () => {
|
|
const usage = { cacheRead: 0, cacheWrite: 0 };
|
|
expect(hasNonzeroUsage(usage)).toBe(false);
|
|
});
|
|
|
|
it("returns false for undefined usage", () => {
|
|
expect(hasNonzeroUsage(undefined)).toBe(false);
|
|
});
|
|
});
|
|
|
|
describe("derivePromptTokens", () => {
|
|
it("includes cache tokens in prompt total", () => {
|
|
const usage = {
|
|
input: 1000,
|
|
cacheRead: 500,
|
|
cacheWrite: 200,
|
|
};
|
|
const promptTokens = derivePromptTokens(usage);
|
|
expect(promptTokens).toBe(1700); // 1000 + 500 + 200
|
|
});
|
|
|
|
it("handles missing cache fields", () => {
|
|
const usage = {
|
|
input: 1000,
|
|
};
|
|
const promptTokens = derivePromptTokens(usage);
|
|
expect(promptTokens).toBe(1000);
|
|
});
|
|
|
|
it("returns undefined for empty usage", () => {
|
|
const promptTokens = derivePromptTokens({});
|
|
expect(promptTokens).toBeUndefined();
|
|
});
|
|
});
|
|
|
|
describe("deriveSessionTotalTokens", () => {
|
|
it("includes cache tokens in total calculation", () => {
|
|
const totalTokens = deriveSessionTotalTokens({
|
|
usage: {
|
|
input: 1000,
|
|
cacheRead: 500,
|
|
cacheWrite: 200,
|
|
},
|
|
contextTokens: 4000,
|
|
});
|
|
expect(totalTokens).toBe(1700); // 1000 + 500 + 200
|
|
});
|
|
|
|
it("prefers promptTokens override over derived total", () => {
|
|
const totalTokens = deriveSessionTotalTokens({
|
|
usage: {
|
|
input: 1000,
|
|
cacheRead: 500,
|
|
cacheWrite: 200,
|
|
},
|
|
contextTokens: 4000,
|
|
promptTokens: 2500, // Override
|
|
});
|
|
expect(totalTokens).toBe(2500);
|
|
});
|
|
});
|