mirror of
https://github.com/openclaw/openclaw.git
synced 2026-03-15 12:00:43 +00:00
383 lines
12 KiB
TypeScript
383 lines
12 KiB
TypeScript
import fs from "node:fs/promises";
|
|
import os from "node:os";
|
|
import path from "node:path";
|
|
import { describe, expect, it } from "vitest";
|
|
import type { SessionEntry } from "../../config/sessions.js";
|
|
import {
|
|
appendHistoryEntry,
|
|
buildHistoryContext,
|
|
buildHistoryContextFromEntries,
|
|
buildHistoryContextFromMap,
|
|
buildPendingHistoryContextFromMap,
|
|
clearHistoryEntriesIfEnabled,
|
|
HISTORY_CONTEXT_MARKER,
|
|
recordPendingHistoryEntryIfEnabled,
|
|
} from "./history.js";
|
|
import {
|
|
DEFAULT_MEMORY_FLUSH_SOFT_TOKENS,
|
|
resolveMemoryFlushContextWindowTokens,
|
|
resolveMemoryFlushSettings,
|
|
shouldRunMemoryFlush,
|
|
} from "./memory-flush.js";
|
|
import { CURRENT_MESSAGE_MARKER } from "./mentions.js";
|
|
import { incrementCompactionCount } from "./session-updates.js";
|
|
|
|
async function seedSessionStore(params: {
|
|
storePath: string;
|
|
sessionKey: string;
|
|
entry: Record<string, unknown>;
|
|
}) {
|
|
await fs.mkdir(path.dirname(params.storePath), { recursive: true });
|
|
await fs.writeFile(
|
|
params.storePath,
|
|
JSON.stringify({ [params.sessionKey]: params.entry }, null, 2),
|
|
"utf-8",
|
|
);
|
|
}
|
|
|
|
async function createCompactionSessionFixture(entry: SessionEntry) {
|
|
const tmp = await fs.mkdtemp(path.join(os.tmpdir(), "openclaw-compact-"));
|
|
const storePath = path.join(tmp, "sessions.json");
|
|
const sessionKey = "main";
|
|
const sessionStore: Record<string, SessionEntry> = { [sessionKey]: entry };
|
|
await seedSessionStore({ storePath, sessionKey, entry });
|
|
return { storePath, sessionKey, sessionStore };
|
|
}
|
|
|
|
describe("history helpers", () => {
|
|
it("returns current message when history is empty", () => {
|
|
const result = buildHistoryContext({
|
|
historyText: " ",
|
|
currentMessage: "hello",
|
|
});
|
|
expect(result).toBe("hello");
|
|
});
|
|
|
|
it("wraps history entries and excludes current by default", () => {
|
|
const result = buildHistoryContextFromEntries({
|
|
entries: [
|
|
{ sender: "A", body: "one" },
|
|
{ sender: "B", body: "two" },
|
|
],
|
|
currentMessage: "current",
|
|
formatEntry: (entry) => `${entry.sender}: ${entry.body}`,
|
|
});
|
|
|
|
expect(result).toContain(HISTORY_CONTEXT_MARKER);
|
|
expect(result).toContain("A: one");
|
|
expect(result).not.toContain("B: two");
|
|
expect(result).toContain(CURRENT_MESSAGE_MARKER);
|
|
expect(result).toContain("current");
|
|
});
|
|
|
|
it("trims history to configured limit", () => {
|
|
const historyMap = new Map<string, { sender: string; body: string }[]>();
|
|
|
|
appendHistoryEntry({
|
|
historyMap,
|
|
historyKey: "group",
|
|
limit: 2,
|
|
entry: { sender: "A", body: "one" },
|
|
});
|
|
appendHistoryEntry({
|
|
historyMap,
|
|
historyKey: "group",
|
|
limit: 2,
|
|
entry: { sender: "B", body: "two" },
|
|
});
|
|
appendHistoryEntry({
|
|
historyMap,
|
|
historyKey: "group",
|
|
limit: 2,
|
|
entry: { sender: "C", body: "three" },
|
|
});
|
|
|
|
expect(historyMap.get("group")?.map((entry) => entry.body)).toEqual(["two", "three"]);
|
|
});
|
|
|
|
it("builds context from map and appends entry", () => {
|
|
const historyMap = new Map<string, { sender: string; body: string }[]>();
|
|
historyMap.set("group", [
|
|
{ sender: "A", body: "one" },
|
|
{ sender: "B", body: "two" },
|
|
]);
|
|
|
|
const result = buildHistoryContextFromMap({
|
|
historyMap,
|
|
historyKey: "group",
|
|
limit: 3,
|
|
entry: { sender: "C", body: "three" },
|
|
currentMessage: "current",
|
|
formatEntry: (entry) => `${entry.sender}: ${entry.body}`,
|
|
});
|
|
|
|
expect(historyMap.get("group")?.map((entry) => entry.body)).toEqual(["one", "two", "three"]);
|
|
expect(result).toContain(HISTORY_CONTEXT_MARKER);
|
|
expect(result).toContain("A: one");
|
|
expect(result).toContain("B: two");
|
|
expect(result).not.toContain("C: three");
|
|
});
|
|
|
|
it("builds context from pending map without appending", () => {
|
|
const historyMap = new Map<string, { sender: string; body: string }[]>();
|
|
historyMap.set("group", [
|
|
{ sender: "A", body: "one" },
|
|
{ sender: "B", body: "two" },
|
|
]);
|
|
|
|
const result = buildPendingHistoryContextFromMap({
|
|
historyMap,
|
|
historyKey: "group",
|
|
limit: 3,
|
|
currentMessage: "current",
|
|
formatEntry: (entry) => `${entry.sender}: ${entry.body}`,
|
|
});
|
|
|
|
expect(historyMap.get("group")?.map((entry) => entry.body)).toEqual(["one", "two"]);
|
|
expect(result).toContain(HISTORY_CONTEXT_MARKER);
|
|
expect(result).toContain("A: one");
|
|
expect(result).toContain("B: two");
|
|
expect(result).toContain(CURRENT_MESSAGE_MARKER);
|
|
expect(result).toContain("current");
|
|
});
|
|
|
|
it("records pending entries only when enabled", () => {
|
|
const historyMap = new Map<string, { sender: string; body: string }[]>();
|
|
|
|
recordPendingHistoryEntryIfEnabled({
|
|
historyMap,
|
|
historyKey: "group",
|
|
limit: 0,
|
|
entry: { sender: "A", body: "one" },
|
|
});
|
|
expect(historyMap.get("group")).toEqual(undefined);
|
|
|
|
recordPendingHistoryEntryIfEnabled({
|
|
historyMap,
|
|
historyKey: "group",
|
|
limit: 2,
|
|
entry: null,
|
|
});
|
|
expect(historyMap.get("group")).toEqual(undefined);
|
|
|
|
recordPendingHistoryEntryIfEnabled({
|
|
historyMap,
|
|
historyKey: "group",
|
|
limit: 2,
|
|
entry: { sender: "B", body: "two" },
|
|
});
|
|
expect(historyMap.get("group")?.map((entry) => entry.body)).toEqual(["two"]);
|
|
});
|
|
|
|
it("clears history entries only when enabled", () => {
|
|
const historyMap = new Map<string, { sender: string; body: string }[]>();
|
|
historyMap.set("group", [
|
|
{ sender: "A", body: "one" },
|
|
{ sender: "B", body: "two" },
|
|
]);
|
|
|
|
clearHistoryEntriesIfEnabled({ historyMap, historyKey: "group", limit: 0 });
|
|
expect(historyMap.get("group")?.map((entry) => entry.body)).toEqual(["one", "two"]);
|
|
|
|
clearHistoryEntriesIfEnabled({ historyMap, historyKey: "group", limit: 2 });
|
|
expect(historyMap.get("group")).toEqual([]);
|
|
});
|
|
});
|
|
|
|
describe("memory flush settings", () => {
|
|
it("defaults to enabled with fallback prompt and system prompt", () => {
|
|
const settings = resolveMemoryFlushSettings();
|
|
expect(settings).not.toBeNull();
|
|
expect(settings?.enabled).toBe(true);
|
|
expect(settings?.prompt.length).toBeGreaterThan(0);
|
|
expect(settings?.systemPrompt.length).toBeGreaterThan(0);
|
|
});
|
|
|
|
it("respects disable flag", () => {
|
|
expect(
|
|
resolveMemoryFlushSettings({
|
|
agents: {
|
|
defaults: { compaction: { memoryFlush: { enabled: false } } },
|
|
},
|
|
}),
|
|
).toBeNull();
|
|
});
|
|
|
|
it("appends NO_REPLY hint when missing", () => {
|
|
const settings = resolveMemoryFlushSettings({
|
|
agents: {
|
|
defaults: {
|
|
compaction: {
|
|
memoryFlush: {
|
|
prompt: "Write memories now.",
|
|
systemPrompt: "Flush memory.",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
});
|
|
expect(settings?.prompt).toContain("NO_REPLY");
|
|
expect(settings?.systemPrompt).toContain("NO_REPLY");
|
|
});
|
|
});
|
|
|
|
describe("shouldRunMemoryFlush", () => {
|
|
it("requires totalTokens and threshold", () => {
|
|
expect(
|
|
shouldRunMemoryFlush({
|
|
entry: { totalTokens: 0 },
|
|
contextWindowTokens: 16_000,
|
|
reserveTokensFloor: 20_000,
|
|
softThresholdTokens: DEFAULT_MEMORY_FLUSH_SOFT_TOKENS,
|
|
}),
|
|
).toBe(false);
|
|
});
|
|
|
|
it("skips when entry is missing", () => {
|
|
expect(
|
|
shouldRunMemoryFlush({
|
|
entry: undefined,
|
|
contextWindowTokens: 16_000,
|
|
reserveTokensFloor: 1_000,
|
|
softThresholdTokens: DEFAULT_MEMORY_FLUSH_SOFT_TOKENS,
|
|
}),
|
|
).toBe(false);
|
|
});
|
|
|
|
it("skips when under threshold", () => {
|
|
expect(
|
|
shouldRunMemoryFlush({
|
|
entry: { totalTokens: 10_000 },
|
|
contextWindowTokens: 100_000,
|
|
reserveTokensFloor: 20_000,
|
|
softThresholdTokens: 10_000,
|
|
}),
|
|
).toBe(false);
|
|
});
|
|
|
|
it("triggers at the threshold boundary", () => {
|
|
expect(
|
|
shouldRunMemoryFlush({
|
|
entry: { totalTokens: 85 },
|
|
contextWindowTokens: 100,
|
|
reserveTokensFloor: 10,
|
|
softThresholdTokens: 5,
|
|
}),
|
|
).toBe(true);
|
|
});
|
|
|
|
it("skips when already flushed for current compaction count", () => {
|
|
expect(
|
|
shouldRunMemoryFlush({
|
|
entry: {
|
|
totalTokens: 90_000,
|
|
compactionCount: 2,
|
|
memoryFlushCompactionCount: 2,
|
|
},
|
|
contextWindowTokens: 100_000,
|
|
reserveTokensFloor: 5_000,
|
|
softThresholdTokens: 2_000,
|
|
}),
|
|
).toBe(false);
|
|
});
|
|
|
|
it("runs when above threshold and not flushed", () => {
|
|
expect(
|
|
shouldRunMemoryFlush({
|
|
entry: { totalTokens: 96_000, compactionCount: 1 },
|
|
contextWindowTokens: 100_000,
|
|
reserveTokensFloor: 5_000,
|
|
softThresholdTokens: 2_000,
|
|
}),
|
|
).toBe(true);
|
|
});
|
|
|
|
it("ignores stale cached totals", () => {
|
|
expect(
|
|
shouldRunMemoryFlush({
|
|
entry: { totalTokens: 96_000, totalTokensFresh: false, compactionCount: 1 },
|
|
contextWindowTokens: 100_000,
|
|
reserveTokensFloor: 5_000,
|
|
softThresholdTokens: 2_000,
|
|
}),
|
|
).toBe(false);
|
|
});
|
|
});
|
|
|
|
describe("resolveMemoryFlushContextWindowTokens", () => {
|
|
it("falls back to agent config or default tokens", () => {
|
|
expect(resolveMemoryFlushContextWindowTokens({ agentCfgContextTokens: 42_000 })).toBe(42_000);
|
|
});
|
|
});
|
|
|
|
describe("incrementCompactionCount", () => {
|
|
it("increments compaction count", async () => {
|
|
const tmp = await fs.mkdtemp(path.join(os.tmpdir(), "openclaw-compact-"));
|
|
const storePath = path.join(tmp, "sessions.json");
|
|
const sessionKey = "main";
|
|
const entry = { sessionId: "s1", updatedAt: Date.now(), compactionCount: 2 } as SessionEntry;
|
|
const sessionStore: Record<string, SessionEntry> = { [sessionKey]: entry };
|
|
await seedSessionStore({ storePath, sessionKey, entry });
|
|
|
|
const count = await incrementCompactionCount({
|
|
sessionEntry: entry,
|
|
sessionStore,
|
|
sessionKey,
|
|
storePath,
|
|
});
|
|
expect(count).toBe(3);
|
|
|
|
const stored = JSON.parse(await fs.readFile(storePath, "utf-8"));
|
|
expect(stored[sessionKey].compactionCount).toBe(3);
|
|
});
|
|
|
|
it("updates totalTokens when tokensAfter is provided", async () => {
|
|
const entry = {
|
|
sessionId: "s1",
|
|
updatedAt: Date.now(),
|
|
compactionCount: 0,
|
|
totalTokens: 180_000,
|
|
inputTokens: 170_000,
|
|
outputTokens: 10_000,
|
|
} as SessionEntry;
|
|
const { storePath, sessionKey, sessionStore } = await createCompactionSessionFixture(entry);
|
|
|
|
await incrementCompactionCount({
|
|
sessionEntry: entry,
|
|
sessionStore,
|
|
sessionKey,
|
|
storePath,
|
|
tokensAfter: 12_000,
|
|
});
|
|
|
|
const stored = JSON.parse(await fs.readFile(storePath, "utf-8"));
|
|
expect(stored[sessionKey].compactionCount).toBe(1);
|
|
expect(stored[sessionKey].totalTokens).toBe(12_000);
|
|
// input/output cleared since we only have the total estimate
|
|
expect(stored[sessionKey].inputTokens).toBeUndefined();
|
|
expect(stored[sessionKey].outputTokens).toBeUndefined();
|
|
});
|
|
|
|
it("does not update totalTokens when tokensAfter is not provided", async () => {
|
|
const entry = {
|
|
sessionId: "s1",
|
|
updatedAt: Date.now(),
|
|
compactionCount: 0,
|
|
totalTokens: 180_000,
|
|
} as SessionEntry;
|
|
const { storePath, sessionKey, sessionStore } = await createCompactionSessionFixture(entry);
|
|
|
|
await incrementCompactionCount({
|
|
sessionEntry: entry,
|
|
sessionStore,
|
|
sessionKey,
|
|
storePath,
|
|
});
|
|
|
|
const stored = JSON.parse(await fs.readFile(storePath, "utf-8"));
|
|
expect(stored[sessionKey].compactionCount).toBe(1);
|
|
// totalTokens unchanged
|
|
expect(stored[sessionKey].totalTokens).toBe(180_000);
|
|
});
|
|
});
|