Files
openclaw/src/agents/model-fallback.test.ts
2026-05-06 10:21:34 +01:00

1944 lines
56 KiB
TypeScript

import crypto from "node:crypto";
import path from "node:path";
import { afterAll, afterEach, beforeAll, beforeEach, describe, expect, it, vi } from "vitest";
import type { OpenClawConfig } from "../config/config.js";
import { resetLogger, setLoggerOverride } from "../logging/logger.js";
import { createWarnLogCapture } from "../logging/test-helpers/warn-log-capture.js";
import {
clearCurrentPluginMetadataSnapshot,
setCurrentPluginMetadataSnapshot,
} from "../plugins/current-plugin-metadata-snapshot.js";
import { loadPluginMetadataSnapshot } from "../plugins/plugin-metadata-snapshot.js";
import { AUTH_STORE_VERSION } from "./auth-profiles/constants.js";
import type { AuthProfileStore } from "./auth-profiles/types.js";
import { FailoverError } from "./failover-error.js";
import { LiveSessionModelSwitchError } from "./live-model-switch-error.js";
import {
FallbackSummaryError,
__testing,
runWithImageModelFallback,
runWithModelFallback,
} from "./model-fallback.js";
import { classifyEmbeddedPiRunResultForModelFallback } from "./pi-embedded-runner/result-fallback-classifier.js";
import type { EmbeddedPiRunResult } from "./pi-embedded-runner/types.js";
import { makeModelFallbackCfg } from "./test-helpers/model-fallback-config-fixture.js";
vi.mock("../infra/file-lock.js", () => ({
withFileLock: async <T>(_filePath: string, _options: unknown, run: () => Promise<T>) => run(),
}));
vi.mock("../plugins/provider-runtime.js", () => ({
buildProviderMissingAuthMessageWithPlugin: () => undefined,
resolveExternalAuthProfilesWithPlugins: () => [],
}));
vi.mock("./provider-model-normalization.runtime.js", () => ({
normalizeProviderModelIdWithRuntime: () => undefined,
}));
const authSourceCheckMock = vi.hoisted(() => ({
hasAnyAuthProfileStoreSource: vi.fn(() => false),
}));
vi.mock("./auth-profiles/source-check.js", () => authSourceCheckMock);
const authRuntimeMock = vi.hoisted(() => {
const stores = new Map<string, AuthProfileStore>();
const keyFor = (agentDir?: string) => agentDir ?? "__main__";
const now = () => Date.now();
const isActive = (value: unknown, ts = now()) =>
typeof value === "number" && Number.isFinite(value) && value > ts;
const getStore = (agentDir?: string): AuthProfileStore =>
stores.get(keyFor(agentDir)) ?? { version: 1, profiles: {} };
const getProfileIds = (store: AuthProfileStore, provider: string) =>
Object.entries(store.profiles)
.filter(([, profile]) => profile.provider === provider)
.map(([id]) => id);
const isProfileInCooldown = (
store: AuthProfileStore,
profileId: string,
tsOrOptions?: number | { now?: number; forModel?: string },
forModel?: string,
) => {
const stats = store.usageStats?.[profileId];
if (!stats || store.profiles[profileId]?.provider === "openrouter") {
return false;
}
const ts = typeof tsOrOptions === "number" ? tsOrOptions : (tsOrOptions?.now ?? now());
const model = typeof tsOrOptions === "object" ? tsOrOptions.forModel : forModel;
if (isActive(stats.disabledUntil, ts)) {
return true;
}
if (!isActive(stats.cooldownUntil, ts)) {
return false;
}
return !stats.cooldownModel || !model || stats.cooldownModel === model;
};
const resolveReason = (store: AuthProfileStore, profileIds: string[], ts = now()) => {
for (const profileId of profileIds) {
const stats = store.usageStats?.[profileId];
if (!stats) {
continue;
}
if (isActive(stats.disabledUntil, ts)) {
return stats.disabledReason ?? "auth";
}
if (!isActive(stats.cooldownUntil, ts)) {
continue;
}
if (stats.cooldownReason) {
return stats.cooldownReason;
}
const counts = stats.failureCounts ?? {};
if ((counts.rate_limit ?? 0) > 0) {
return "rate_limit";
}
if ((counts.overloaded ?? 0) > 0) {
return "overloaded";
}
if ((counts.timeout ?? 0) > 0) {
return "timeout";
}
return "unknown";
}
return null;
};
return {
clear: () => stores.clear(),
setStore: (agentDir: string | undefined, store: AuthProfileStore) => {
stores.set(keyFor(agentDir), store);
},
runtime: {
ensureAuthProfileStore: vi.fn((agentDir?: string) => getStore(agentDir)),
loadAuthProfileStoreForRuntime: vi.fn((agentDir?: string) => getStore(agentDir)),
resolveAuthProfileOrder: (params: { store: AuthProfileStore; provider: string }) =>
getProfileIds(params.store, params.provider),
isProfileInCooldown,
resolveProfilesUnavailableReason: (params: {
store: AuthProfileStore;
profileIds: string[];
now?: number;
}) => resolveReason(params.store, params.profileIds, params.now),
getSoonestCooldownExpiry: (
store: AuthProfileStore,
profileIds: string[],
options?: { now?: number; forModel?: string },
) => {
const ts = options?.now ?? now();
let soonest: number | null = null;
for (const profileId of profileIds) {
if (!isProfileInCooldown(store, profileId, { now: ts, forModel: options?.forModel })) {
continue;
}
const stats = store.usageStats?.[profileId];
const expiry = [stats?.cooldownUntil, stats?.disabledUntil]
.filter((value): value is number => isActive(value, ts))
.toSorted((a, b) => a - b)[0];
if (expiry !== undefined && (soonest === null || expiry < soonest)) {
soonest = expiry;
}
}
return soonest;
},
},
};
});
vi.mock("./model-fallback-auth.runtime.js", () => authRuntimeMock.runtime);
const makeCfg = makeModelFallbackCfg;
let authTempRoot = "";
let authTempCounter = 0;
beforeAll(() => {
setCurrentPluginMetadataSnapshot(loadPluginMetadataSnapshot({ config: {}, env: process.env }), {
config: {},
env: process.env,
});
});
afterAll(() => {
clearCurrentPluginMetadataSnapshot();
});
function resetModelFallbackTestState(): void {
authRuntimeMock.clear();
authRuntimeMock.runtime.ensureAuthProfileStore.mockClear();
authRuntimeMock.runtime.loadAuthProfileStoreForRuntime.mockClear();
authSourceCheckMock.hasAnyAuthProfileStoreSource.mockReset().mockReturnValue(false);
}
afterEach(resetModelFallbackTestState);
beforeEach(() => {
setLoggerOverride({ level: "silent", consoleLevel: "silent" });
});
afterEach(() => {
setLoggerOverride(null);
resetLogger();
});
async function runModelFallbackCase(name: string, run: () => Promise<void>): Promise<void> {
try {
await run();
} catch (err) {
throw new Error(`case failed: ${name}`, { cause: err });
} finally {
resetModelFallbackTestState();
}
}
function makeFallbacksOnlyCfg(): OpenClawConfig {
return {
agents: {
defaults: {
model: {
fallbacks: ["openai/gpt-5.2"],
},
},
},
} as OpenClawConfig;
}
function makeProviderFallbackCfg(provider: string): OpenClawConfig {
return makeCfg({
agents: {
defaults: {
model: {
primary: `${provider}/m1`,
fallbacks: ["fallback/ok-model"],
},
},
},
});
}
async function withTempAuthStore<T>(
store: AuthProfileStore,
run: (tempDir: string) => Promise<T>,
): Promise<T> {
const tempDir = await makeAuthTempDir();
setAuthRuntimeStore(tempDir, store);
return await run(tempDir);
}
async function makeAuthTempDir(): Promise<string> {
authTempRoot ||= path.join("/tmp", "openclaw-auth-suite-mock");
return path.join(authTempRoot, `case-${++authTempCounter}`);
}
async function runWithStoredAuth(params: {
cfg: OpenClawConfig;
store: AuthProfileStore;
provider: string;
run: (provider: string, model: string) => Promise<string>;
}) {
const tempDir = await makeAuthTempDir();
setAuthRuntimeStore(tempDir, params.store);
return await runWithModelFallback({
cfg: params.cfg,
provider: params.provider,
model: "m1",
agentDir: tempDir,
run: params.run,
});
}
function setAuthRuntimeStore(agentDir: string | undefined, store: AuthProfileStore): void {
authSourceCheckMock.hasAnyAuthProfileStoreSource.mockReturnValue(true);
authRuntimeMock.setStore(agentDir, store);
}
async function expectFallsBackToHaiku(params: {
provider: string;
model: string;
firstError: Error;
}) {
const cfg = makeCfg();
const run = vi.fn().mockRejectedValueOnce(params.firstError).mockResolvedValueOnce("ok");
const result = await runWithModelFallback({
cfg,
provider: params.provider,
model: params.model,
run,
});
expect(result.result).toBe("ok");
expect(run).toHaveBeenCalledTimes(2);
expect(run.mock.calls[1]?.[0]).toBe("anthropic");
expect(run.mock.calls[1]?.[1]).toBe("claude-haiku-3-5");
}
function createOverrideFailureRun(params: {
overrideProvider: string;
overrideModel: string;
fallbackProvider: string;
fallbackModel: string;
firstError: Error;
}) {
return vi.fn().mockImplementation(async (provider, model) => {
if (provider === params.overrideProvider && model === params.overrideModel) {
throw params.firstError;
}
if (provider === params.fallbackProvider && model === params.fallbackModel) {
return "ok";
}
throw new Error(`unexpected fallback candidate: ${provider}/${model}`);
});
}
function makeSingleProviderStore(params: {
provider: string;
usageStat: NonNullable<AuthProfileStore["usageStats"]>[string];
}): AuthProfileStore {
const profileId = `${params.provider}:default`;
return {
version: AUTH_STORE_VERSION,
profiles: {
[profileId]: {
type: "api_key",
provider: params.provider,
key: "test-key",
},
},
usageStats: {
[profileId]: params.usageStat,
},
};
}
function createFallbackOnlyRun() {
return vi.fn().mockImplementation(async (providerId, modelId) => {
if (providerId === "fallback") {
return "ok";
}
throw new Error(`unexpected provider: ${providerId}/${modelId}`);
});
}
async function expectSkippedUnavailableProvider(params: {
providerPrefix: string;
usageStat: NonNullable<AuthProfileStore["usageStats"]>[string];
expectedReason: string;
}) {
const provider = `${params.providerPrefix}-${crypto.randomUUID()}`;
const cfg = makeProviderFallbackCfg(provider);
const primaryStore = makeSingleProviderStore({
provider,
usageStat: params.usageStat,
});
// Include fallback provider profile so the fallback is attempted (not skipped as no-profile).
const store: AuthProfileStore = {
...primaryStore,
profiles: {
...primaryStore.profiles,
"fallback:default": {
type: "api_key",
provider: "fallback",
key: "test-key",
},
},
};
const run = createFallbackOnlyRun();
const result = await runWithStoredAuth({
cfg,
store,
provider,
run,
});
expect(result.result).toBe("ok");
expect(run.mock.calls).toEqual([["fallback", "ok-model"]]);
expect(result.attempts[0]?.reason).toBe(params.expectedReason);
}
// Issue-backed Anthropic/OpenAI-compatible insufficient_quota payload under HTTP 400:
// https://github.com/openclaw/openclaw/issues/23440
const INSUFFICIENT_QUOTA_PAYLOAD =
'{"type":"error","error":{"type":"insufficient_quota","message":"Your account has insufficient quota balance to run this request."}}';
describe("runWithModelFallback", () => {
it("skips auth store bootstrap when no auth profile sources exist", async () => {
authSourceCheckMock.hasAnyAuthProfileStoreSource.mockReturnValue(false);
const run = vi.fn().mockResolvedValueOnce("ok");
const result = await runWithModelFallback({
cfg: makeCfg(),
provider: "openai",
model: "gpt-4.1-mini",
agentDir: "/tmp/openclaw-no-auth-profiles",
run,
});
expect(result.result).toBe("ok");
expect(authSourceCheckMock.hasAnyAuthProfileStoreSource).toHaveBeenCalledWith(
"/tmp/openclaw-no-auth-profiles",
);
expect(authRuntimeMock.runtime.ensureAuthProfileStore).not.toHaveBeenCalled();
expect(run).toHaveBeenCalledWith("openai", "gpt-4.1-mini");
});
it("resolves primary model aliases before running", () => {
const cases = [
{
name: "keeps openai gpt-5.4 on provider",
cfg: makeCfg(),
provider: "openai",
model: "gpt-5.4",
expected: ["openai", "gpt-5.4"],
},
{
name: "resolves bare alias",
cfg: makeCfg({
agents: {
defaults: {
model: {
primary: "anthropic/claude-sonnet-4-6",
fallbacks: [],
},
models: {
"anthropic/claude-sonnet-4-6": { alias: "sonnet" },
},
},
},
}),
provider: "anthropic",
model: "sonnet",
expected: ["anthropic", "claude-sonnet-4-6"],
},
{
name: "resolves slash-form alias before provider parsing",
cfg: makeCfg({
agents: {
defaults: {
model: {
primary: "openai/xiaomi/mimo-v2-pro-mit",
fallbacks: [],
},
models: {
"openai/xiaomi/mimo-v2-pro-mit": { alias: "xiaomi/mimo-v2-pro-mit" },
},
},
},
}),
provider: "xiaomi",
model: "mimo-v2-pro-mit",
expected: ["openai", "xiaomi/mimo-v2-pro-mit"],
},
] satisfies Array<{
name: string;
cfg: OpenClawConfig;
provider: string;
model: string;
expected: [string, string];
}>;
for (const testCase of cases) {
const candidates = __testing.resolveFallbackCandidates({
cfg: testCase.cfg,
provider: testCase.provider,
model: testCase.model,
});
expect(candidates[0], testCase.name).toEqual({
provider: testCase.expected[0],
model: testCase.expected[1],
});
}
});
it("falls back on unrecognized errors when candidates remain", async () => {
const cfg = makeCfg();
const run = vi.fn().mockRejectedValueOnce(new Error("bad request")).mockResolvedValueOnce("ok");
const result = await runWithModelFallback({
cfg,
provider: "openai",
model: "gpt-4.1-mini",
run,
});
expect(result.result).toBe("ok");
expect(run).toHaveBeenCalledTimes(2);
expect(result.attempts).toHaveLength(1);
expect(result.attempts[0].error).toBe("bad request");
expect(result.attempts[0].reason).toBe("unknown");
});
it("keeps raw provider schema errors in fallback summaries", async () => {
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "openai/gpt-5.4",
fallbacks: ["openai/gpt-5.4-mini"],
},
},
},
});
const rawError =
"400 The following tools cannot be used with reasoning.effort 'minimal': web_search.";
const run = vi.fn().mockRejectedValue(
new FailoverError("LLM request failed: provider rejected the request schema.", {
provider: "openai",
model: "gpt-5.4",
reason: "format",
status: 400,
rawError,
}),
);
await expect(
runWithModelFallback({
cfg,
provider: "openai",
model: "gpt-5.4",
run,
}),
).rejects.toMatchObject({
name: "FallbackSummaryError",
message: expect.stringContaining(rawError),
attempts: expect.arrayContaining([
expect.objectContaining({
error: rawError,
reason: "format",
status: 400,
}),
]),
});
});
it("carries request attribution through exhausted fallback summaries", async () => {
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "openai/gpt-5.4",
fallbacks: ["anthropic/claude-opus-4-6"],
},
},
},
});
const run = vi
.fn()
.mockRejectedValueOnce(Object.assign(new Error("rate limit exceeded"), { status: 429 }))
.mockRejectedValueOnce(Object.assign(new Error("overloaded"), { status: 503 }));
try {
await runWithModelFallback({
cfg,
provider: "openai",
model: "gpt-5.4",
runId: "run-42713",
sessionId: "session:browser-42713",
lane: "answer",
run,
});
throw new Error("expected fallback summary");
} catch (err) {
expect(err).toBeInstanceOf(FallbackSummaryError);
if (!(err instanceof FallbackSummaryError)) {
throw err;
}
expect(err).toMatchObject({
name: "FallbackSummaryError",
sessionId: "session:browser-42713",
lane: "answer",
});
expect(err.cause).toMatchObject({
name: "FailoverError",
sessionId: "session:browser-42713",
lane: "answer",
});
}
});
it("uses optional result classification to continue to configured fallbacks", async () => {
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "openai-codex/gpt-5.4",
fallbacks: ["anthropic/claude-haiku-3-5"],
},
},
},
});
const run = vi
.fn()
.mockResolvedValueOnce({ payloads: [] })
.mockResolvedValueOnce({
payloads: [{ text: "fallback ok" }],
});
const classifyResult = vi.fn(({ result }) =>
Array.isArray(result.payloads) && result.payloads.length === 0
? {
message: "terminal result contained no visible assistant reply",
reason: "format" as const,
code: "empty_result",
}
: null,
);
const result = await runWithModelFallback({
cfg,
provider: "openai-codex",
model: "gpt-5.4",
run,
classifyResult,
});
expect(result.result).toEqual({ payloads: [{ text: "fallback ok" }] });
expect(run).toHaveBeenCalledTimes(2);
expect(run.mock.calls[1]).toEqual(["anthropic", "claude-haiku-3-5"]);
expect(result.attempts[0]).toMatchObject({
provider: "openai-codex",
model: "gpt-5.4",
reason: "format",
code: "empty_result",
});
});
it("surfaces classified terminal results when no fallback remains", async () => {
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "openai-codex/gpt-5.4",
fallbacks: [],
},
},
},
});
const run = vi.fn().mockResolvedValueOnce({ payloads: [] });
await expect(
runWithModelFallback({
cfg,
provider: "openai-codex",
model: "gpt-5.4",
run,
classifyResult: ({ result }) => {
const payloads = (result as { payloads?: unknown[] }).payloads;
return Array.isArray(payloads) && payloads.length === 0
? {
message: "terminal result contained no visible assistant reply",
reason: "format",
code: "empty_result",
}
: null;
},
}),
).rejects.toMatchObject({
name: "FailoverError",
reason: "format",
provider: "openai-codex",
model: "gpt-5.4",
code: "empty_result",
});
expect(run).toHaveBeenCalledTimes(1);
});
it("does not classify successful results when the optional classifier returns null", async () => {
const cfg = makeProviderFallbackCfg("openai-codex");
const run = vi.fn().mockResolvedValueOnce({ payloads: [{ text: "ok" }] });
const classifyResult = vi.fn(() => null);
const result = await runWithModelFallback({
cfg,
provider: "openai-codex",
model: "m1",
run,
classifyResult,
});
expect(result.result).toEqual({ payloads: [{ text: "ok" }] });
expect(run).toHaveBeenCalledTimes(1);
expect(result.attempts).toEqual([]);
});
it("keeps tool-executing empty GPT-5 runs out of fallback", () => {
const runResult: EmbeddedPiRunResult = {
payloads: [],
meta: {
durationMs: 1,
toolSummary: {
calls: 1,
tools: ["mcp_write"],
},
},
};
expect(
classifyEmbeddedPiRunResultForModelFallback({
provider: "openai-codex",
model: "gpt-5.4",
result: runResult,
}),
).toBeNull();
});
it("keeps normalized silent GPT-5 terminal replies out of fallback", () => {
const runResult: EmbeddedPiRunResult = {
payloads: [],
meta: {
durationMs: 1,
finalAssistantRawText: "NO_REPLY",
},
};
expect(
classifyEmbeddedPiRunResultForModelFallback({
provider: "openai-codex",
model: "gpt-5.4",
result: runResult,
}),
).toBeNull();
});
it("uses harness-owned terminal classification for GPT-5 fallback", () => {
const runResult: EmbeddedPiRunResult = {
payloads: [],
meta: {
durationMs: 1,
agentHarnessResultClassification: "planning-only",
},
};
expect(
classifyEmbeddedPiRunResultForModelFallback({
provider: "codex",
model: "gpt-5.4",
result: runResult,
}),
).toMatchObject({
code: "planning_only_result",
reason: "format",
});
});
it("classifies non-GPT incomplete terminal errors for configured fallback", () => {
const runResult: EmbeddedPiRunResult = {
payloads: [
{ text: "⚠️ Agent couldn't generate a response. Please try again.", isError: true },
],
meta: {
durationMs: 1,
},
};
expect(
classifyEmbeddedPiRunResultForModelFallback({
provider: "anthropic",
model: "claude-opus-4.7",
result: runResult,
}),
).toMatchObject({
code: "incomplete_result",
reason: "format",
});
});
it("keeps aborted harness-classified GPT-5 runs out of fallback", () => {
const runResult: EmbeddedPiRunResult = {
payloads: [],
meta: {
durationMs: 1,
aborted: true,
agentHarnessResultClassification: "empty",
},
};
expect(
classifyEmbeddedPiRunResultForModelFallback({
provider: "codex",
model: "gpt-5.4",
result: runResult,
}),
).toBeNull();
});
it("passes original unknown errors to onError during fallback", async () => {
const cfg = makeCfg();
const unknownError = new Error("provider misbehaved");
const run = vi.fn().mockRejectedValueOnce(unknownError).mockResolvedValueOnce("ok");
const onError = vi.fn();
await runWithModelFallback({
cfg,
provider: "openai",
model: "gpt-4.1-mini",
run,
onError,
});
expect(onError).toHaveBeenCalledTimes(1);
expect(onError.mock.calls[0]?.[0]).toMatchObject({
provider: "openai",
model: "gpt-4.1-mini",
attempt: 1,
total: 2,
});
expect(onError.mock.calls[0]?.[0]?.error).toBe(unknownError);
});
it("throws unrecognized error on last candidate", async () => {
const cfg = makeCfg();
const run = vi.fn().mockRejectedValueOnce(new Error("something weird"));
await expect(
runWithModelFallback({
cfg,
provider: "openai",
model: "gpt-4.1-mini",
run,
fallbacksOverride: [],
}),
).rejects.toThrow("something weird");
expect(run).toHaveBeenCalledTimes(1);
});
it("treats LiveSessionModelSwitchError as failover on last candidate (#58496 family)", async () => {
const cfg = makeCfg();
const switchError = new LiveSessionModelSwitchError({
provider: "anthropic",
model: "claude-sonnet-4-6",
});
const run = vi.fn().mockRejectedValue(switchError);
// With no fallbacks, the single candidate is also the last one.
// Previously this would re-throw LiveSessionModelSwitchError, causing
// the outer retry loop to restart with the overloaded model indefinitely.
// Now it should surface as a FailoverError instead.
const err = await runWithModelFallback({
cfg,
provider: "anthropic",
model: "claude-sonnet-4-6",
run,
fallbacksOverride: [],
}).catch((e: unknown) => e);
expect(err).toBeInstanceOf(Error);
// Should NOT be a LiveSessionModelSwitchError — the outer retry loop must
// not restart with the conflicting model.
expect(err).not.toBeInstanceOf(LiveSessionModelSwitchError);
expect((err as { reason?: string }).reason).toBe("unknown");
expect(run).toHaveBeenCalledTimes(1);
});
it("continues fallback chain past LiveSessionModelSwitchError to next candidate (#58496 family)", async () => {
const cfg = makeCfg();
const switchError = new LiveSessionModelSwitchError({
provider: "anthropic",
model: "claude-sonnet-4-6",
});
const run = vi.fn().mockRejectedValueOnce(switchError).mockResolvedValueOnce("ok");
const result = await runWithModelFallback({
cfg,
provider: "openai",
model: "gpt-4.1-mini",
run,
});
expect(result.result).toBe("ok");
expect(run).toHaveBeenCalledTimes(2);
});
it("jumps directly to a later live-session model switch candidate (#57471)", async () => {
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "openai/gpt-4.1-mini",
fallbacks: [
"anthropic/claude-haiku-3-5",
"anthropic/claude-sonnet-4-6",
"openrouter/deepseek-chat",
],
},
},
},
});
const switchError = new LiveSessionModelSwitchError({
provider: "anthropic",
model: "claude-sonnet-4-6",
});
const run = vi.fn(async (provider: string, model: string) => {
if (provider === "openai" && model === "gpt-4.1-mini") {
throw switchError;
}
if (provider === "anthropic" && model === "claude-sonnet-4-6") {
return "ok";
}
throw new Error(`unexpected fallback candidate: ${provider}/${model}`);
});
const onError = vi.fn();
const result = await runWithModelFallback({
cfg,
provider: "openai",
model: "gpt-4.1-mini",
run,
onError,
});
expect(result.result).toBe("ok");
expect(result.provider).toBe("anthropic");
expect(result.model).toBe("claude-sonnet-4-6");
expect(result.attempts).toEqual([]);
expect(onError).not.toHaveBeenCalled();
expect(run.mock.calls).toEqual([
["openai", "gpt-4.1-mini"],
["anthropic", "claude-sonnet-4-6"],
]);
});
it("does not redirect stale live-session switch errors back to the current candidate (#58496 family)", async () => {
const cfg = makeCfg();
const switchError = new LiveSessionModelSwitchError({
provider: "openai",
model: "gpt-4.1-mini",
});
const run = vi.fn().mockRejectedValueOnce(switchError).mockResolvedValueOnce("ok");
const result = await runWithModelFallback({
cfg,
provider: "openai",
model: "gpt-4.1-mini",
run,
});
expect(result.result).toBe("ok");
expect(result.provider).toBe("anthropic");
expect(result.model).toBe("claude-haiku-3-5");
expect(result.attempts[0]?.reason).toBe("unknown");
expect(run.mock.calls).toEqual([
["openai", "gpt-4.1-mini"],
["anthropic", "claude-haiku-3-5"],
]);
});
it("falls back to the configured haiku candidate for retryable provider failures", async () => {
await expectFallsBackToHaiku({
provider: "openai",
model: "gpt-4.1-mini",
firstError: Object.assign(new Error("nope"), { status: 401 }),
});
});
it("puts configured primary next when an override model is requested", () => {
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "openai/gpt-4.1-mini",
fallbacks: ["anthropic/claude-haiku-3-5", "openrouter/deepseek-chat"],
},
},
},
});
expect(
__testing.resolveFallbackCandidates({
cfg,
provider: "anthropic",
model: "claude-opus-4-5",
}),
).toEqual([
{ provider: "anthropic", model: "claude-opus-4-5" },
{ provider: "openai", model: "gpt-4.1-mini" },
]);
});
it("keeps configured fallback chain when current model is a configured fallback", () => {
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "openai/gpt-4.1-mini",
fallbacks: ["anthropic/claude-haiku-3-5", "openrouter/deepseek-chat"],
},
},
},
});
expect(
__testing.resolveFallbackCandidates({
cfg,
provider: "anthropic",
model: "claude-haiku-3-5",
}),
).toEqual([
{ provider: "anthropic", model: "claude-haiku-3-5" },
{ provider: "openrouter", model: "openrouter/deepseek-chat" },
{ provider: "openai", model: "gpt-4.1-mini" },
]);
});
it("treats normalized default refs as primary and keeps configured fallback chain", () => {
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "openai/gpt-4.1-mini",
fallbacks: ["anthropic/claude-haiku-3-5"],
},
},
},
});
expect(
__testing.resolveFallbackCandidates({
cfg,
provider: " OpenAI ",
model: "gpt-4.1-mini",
}),
).toEqual([
{ provider: "openai", model: "gpt-4.1-mini" },
{ provider: "anthropic", model: "claude-haiku-3-5" },
]);
});
it("falls back to configured primary for override credential validation errors", async () => {
const cfg = makeCfg();
const run = createOverrideFailureRun({
overrideProvider: "anthropic",
overrideModel: "claude-opus-4",
fallbackProvider: "openai",
fallbackModel: "gpt-4.1-mini",
firstError: new Error('No credentials found for profile "anthropic:default".'),
});
const result = await runWithModelFallback({
cfg,
provider: "anthropic",
model: "claude-opus-4",
run,
});
expect(result.result).toBe("ok");
expect(run.mock.calls).toEqual([
["anthropic", "claude-opus-4"],
["openai", "gpt-4.1-mini"],
]);
});
it("records 400 insufficient_quota payloads as billing during fallback", async () => {
const cfg = makeCfg();
const run = vi
.fn()
.mockRejectedValueOnce(Object.assign(new Error(INSUFFICIENT_QUOTA_PAYLOAD), { status: 400 }))
.mockResolvedValueOnce("ok");
const result = await runWithModelFallback({
cfg,
provider: "openai",
model: "gpt-4.1-mini",
run,
});
expect(result.result).toBe("ok");
expect(result.attempts).toHaveLength(1);
expect(result.attempts[0]?.reason).toBe("billing");
});
it("falls back on model-not-found error shapes", async () => {
const cases: Array<{
name: string;
provider: string;
model: string;
error: Error;
expectedFallback: [string, string];
expectedReason?: string;
}> = [
{
name: "unknown anthropic override",
provider: "anthropic",
model: "claude-opus-4-6",
error: new Error("Unknown model: anthropic/claude-opus-4-6"),
expectedFallback: ["openai", "gpt-4.1-mini"],
},
{
name: "openai model not found",
provider: "openai",
model: "gpt-6",
error: new Error("Model not found: openai/gpt-6"),
expectedFallback: ["anthropic", "claude-haiku-3-5"],
},
];
for (const testCase of cases) {
await runModelFallbackCase(testCase.name, async () => {
const cfg = makeCfg();
const run = vi.fn().mockRejectedValueOnce(testCase.error).mockResolvedValueOnce("ok");
const result = await runWithModelFallback({
cfg,
provider: testCase.provider,
model: testCase.model,
run,
});
expect(result.result).toBe("ok");
expect(run).toHaveBeenCalledTimes(2);
expect(run.mock.calls[1]).toEqual(testCase.expectedFallback);
if (testCase.expectedReason) {
expect(result.attempts).toHaveLength(1);
expect(result.attempts[0]?.reason).toBe(testCase.expectedReason);
}
});
}
});
it("warns when falling back due to model_not_found", async () => {
setLoggerOverride({ level: "silent", consoleLevel: "warn" });
const warnSpy = vi.spyOn(console, "warn").mockImplementation(() => {});
try {
const cfg = makeCfg();
const run = vi
.fn()
.mockRejectedValueOnce(new Error("Model not found: openai/gpt-6"))
.mockResolvedValueOnce("ok");
const result = await runWithModelFallback({
cfg,
provider: "openai",
model: "gpt-6",
run,
});
expect(result.result).toBe("ok");
expect(warnSpy).toHaveBeenCalledWith(
expect.stringContaining('Model "openai/gpt-6" not found'),
);
} finally {
warnSpy.mockRestore();
setLoggerOverride(null);
resetLogger();
}
});
it("sanitizes model identifiers in model_not_found warnings", async () => {
const warnLogs = createWarnLogCapture("openclaw-model-fallback-test");
try {
const cfg = makeCfg();
const run = vi
.fn()
.mockRejectedValueOnce(new Error("Model not found: openai/gpt-6"))
.mockResolvedValueOnce("ok");
const result = await runWithModelFallback({
cfg,
provider: "openai",
model: "gpt-6\u001B[31m\nspoof",
run,
});
expect(result.result).toBe("ok");
const warning = await warnLogs.findText('Model "openai/gpt-6spoof" not found');
expect(warning).toContain('Model "openai/gpt-6spoof" not found');
expect(warning).not.toContain("\u001B");
expect(warning).not.toContain("\n");
} finally {
warnLogs.cleanup();
}
});
it("skips providers when all profiles are in cooldown", async () => {
await expectSkippedUnavailableProvider({
providerPrefix: "cooldown-test",
usageStat: {
cooldownUntil: Date.now() + 5 * 60_000,
},
expectedReason: "unknown",
});
});
it("does not skip OpenRouter when legacy cooldown markers exist", async () => {
const provider = "openrouter";
const cfg = makeProviderFallbackCfg(provider);
const store = makeSingleProviderStore({
provider,
usageStat: {
cooldownUntil: Date.now() + 5 * 60_000,
disabledUntil: Date.now() + 10 * 60_000,
disabledReason: "billing",
},
});
const run = vi.fn().mockImplementation(async (providerId) => {
if (providerId === "openrouter") {
return "ok";
}
throw new Error(`unexpected provider: ${providerId}`);
});
const result = await runWithStoredAuth({
cfg,
store,
provider,
run,
});
expect(result.result).toBe("ok");
expect(run).toHaveBeenCalledTimes(1);
expect(run.mock.calls[0]?.[0]).toBe("openrouter");
expect(result.attempts).toEqual([]);
});
it("propagates disabled reason when all profiles are unavailable", async () => {
const now = Date.now();
await expectSkippedUnavailableProvider({
providerPrefix: "disabled-test",
usageStat: {
disabledUntil: now + 5 * 60_000,
disabledReason: "billing",
failureCounts: { rate_limit: 4 },
},
expectedReason: "billing",
});
});
it("does not skip when any profile is available", async () => {
const provider = `cooldown-mixed-${crypto.randomUUID()}`;
const profileA = `${provider}:a`;
const profileB = `${provider}:b`;
const store: AuthProfileStore = {
version: AUTH_STORE_VERSION,
profiles: {
[profileA]: {
type: "api_key",
provider,
key: "key-a",
},
[profileB]: {
type: "api_key",
provider,
key: "key-b",
},
},
usageStats: {
[profileA]: {
cooldownUntil: Date.now() + 60_000,
},
},
};
const cfg = makeProviderFallbackCfg(provider);
const run = vi.fn().mockImplementation(async (providerId) => {
if (providerId === provider) {
return "ok";
}
return "unexpected";
});
const result = await runWithStoredAuth({
cfg,
store,
provider,
run,
});
expect(result.result).toBe("ok");
expect(run.mock.calls).toEqual([[provider, "m1"]]);
expect(result.attempts).toEqual([]);
});
it("does not append configured primary when fallbacksOverride is set", () => {
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "openai/gpt-4.1-mini",
},
},
},
});
expect(
__testing.resolveFallbackCandidates({
cfg,
provider: "anthropic",
model: "claude-opus-4-5",
fallbacksOverride: ["anthropic/claude-haiku-3-5"],
}),
).toEqual([
{ provider: "anthropic", model: "claude-opus-4-5" },
{ provider: "anthropic", model: "claude-haiku-3-5" },
]);
});
it("refreshes cooldown expiry from persisted auth state before fallback summary", async () => {
const expiry = Date.now() + 120_000;
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "anthropic/claude-opus-4-5",
fallbacks: ["openai/gpt-5.2"],
},
},
},
});
const store: AuthProfileStore = {
version: AUTH_STORE_VERSION,
profiles: {
"anthropic:default": { type: "api_key", provider: "anthropic", key: "anthropic-key" },
"openai:default": { type: "api_key", provider: "openai", key: "openai-key" },
},
};
await withTempAuthStore(store, async (tempDir) => {
const run = vi.fn().mockImplementation(async (provider: string, model: string) => {
if (provider === "anthropic" && model === "claude-opus-4-5") {
setAuthRuntimeStore(tempDir, {
...store,
usageStats: {
"anthropic:default": {
cooldownUntil: expiry,
cooldownReason: "rate_limit",
cooldownModel: "claude-opus-4-5",
failureCounts: { rate_limit: 1 },
},
},
});
}
throw Object.assign(new Error("rate limited"), { status: 429 });
});
await expect(
runWithModelFallback({
cfg,
provider: "anthropic",
model: "claude-opus-4-5",
agentDir: tempDir,
run,
}),
).rejects.toMatchObject({
name: "FallbackSummaryError",
soonestCooldownExpiry: expiry,
});
});
});
it("filters fallback summary cooldown expiry to attempted model scopes", async () => {
const now = Date.now();
const unrelatedExpiry = now + 15_000;
const relevantExpiry = now + 90_000;
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "anthropic/claude-opus-4-5",
fallbacks: ["openai/gpt-5.2"],
},
},
},
});
const store: AuthProfileStore = {
version: AUTH_STORE_VERSION,
profiles: {
"anthropic:default": { type: "api_key", provider: "anthropic", key: "anthropic-key" },
"openai:default": { type: "api_key", provider: "openai", key: "openai-key" },
},
usageStats: {
"anthropic:default": {
cooldownUntil: unrelatedExpiry,
cooldownReason: "rate_limit",
cooldownModel: "claude-haiku-3-5",
failureCounts: { rate_limit: 1 },
},
"openai:default": {
cooldownUntil: relevantExpiry,
cooldownReason: "rate_limit",
cooldownModel: "gpt-5.2",
failureCounts: { rate_limit: 1 },
},
},
};
await withTempAuthStore(store, async (tempDir) => {
const run = vi
.fn()
.mockRejectedValue(Object.assign(new Error("rate limited"), { status: 429 }));
await expect(
runWithModelFallback({
cfg,
provider: "anthropic",
model: "claude-opus-4-5",
agentDir: tempDir,
run,
}),
).rejects.toMatchObject({
name: "FallbackSummaryError",
soonestCooldownExpiry: relevantExpiry,
});
});
});
it("uses fallbacksOverride instead of agents.defaults.model.fallbacks", async () => {
const cfg = makeFallbacksOnlyCfg();
const candidates = __testing.resolveFallbackCandidates({
cfg,
provider: "anthropic",
model: "claude-opus-4-5",
fallbacksOverride: ["openai/gpt-4.1"],
});
expect(candidates).toEqual([
{ provider: "anthropic", model: "claude-opus-4-5" },
{ provider: "openai", model: "gpt-4.1" },
]);
});
it("treats an empty fallbacksOverride as disabling global fallbacks", async () => {
const cfg = makeFallbacksOnlyCfg();
const candidates = __testing.resolveFallbackCandidates({
cfg,
provider: "anthropic",
model: "claude-opus-4-5",
fallbacksOverride: [],
});
expect(candidates).toEqual([{ provider: "anthropic", model: "claude-opus-4-5" }]);
});
it("keeps explicit fallbacks reachable when models allowlist is present", async () => {
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "anthropic/claude-sonnet-4",
fallbacks: ["openai/gpt-4o", "ollama/llama-3"],
},
models: {
"anthropic/claude-sonnet-4": {},
},
},
},
});
const candidates = __testing.resolveFallbackCandidates({
cfg,
provider: "anthropic",
model: "claude-sonnet-4",
});
expect(candidates).toEqual([
{ provider: "anthropic", model: "claude-sonnet-4" },
{ provider: "openai", model: "gpt-4o" },
{ provider: "ollama", model: "llama-3" },
]);
});
it("defaults provider/model when missing (regression #946)", async () => {
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "openai/gpt-4.1-mini",
fallbacks: [],
},
},
},
});
const candidates = __testing.resolveFallbackCandidates({
cfg,
provider: undefined as unknown as string,
model: undefined as unknown as string,
});
expect(candidates).toEqual([{ provider: "openai", model: "gpt-4.1-mini" }]);
});
it("does not fall back on user aborts", async () => {
const cfg = makeCfg();
const run = vi
.fn()
.mockRejectedValueOnce(Object.assign(new Error("aborted"), { name: "AbortError" }))
.mockResolvedValueOnce("ok");
await expect(
runWithModelFallback({
cfg,
provider: "openai",
model: "gpt-4.1-mini",
run,
}),
).rejects.toThrow("aborted");
expect(run).toHaveBeenCalledTimes(1);
});
it("appends the configured primary as a last fallback", async () => {
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "openai/gpt-4.1-mini",
fallbacks: [],
},
},
},
});
const run = vi
.fn()
.mockRejectedValueOnce(Object.assign(new Error("timeout"), { code: "ETIMEDOUT" }))
.mockResolvedValueOnce("ok");
const result = await runWithModelFallback({
cfg,
provider: "openrouter",
model: "meta-llama/llama-3.3-70b:free",
run,
});
expect(result.result).toBe("ok");
expect(run).toHaveBeenCalledTimes(2);
expect(result.provider).toBe("openai");
expect(result.model).toBe("gpt-4.1-mini");
});
// Tests for Bug A fix: Model fallback with session overrides
describe("fallback behavior with session model overrides", () => {
it("keeps fallback ordering correct across session overrides", () => {
const cases = [
{
name: "same provider versioned session model",
cfg: makeCfg({
agents: {
defaults: {
model: {
primary: "anthropic/claude-opus-4-6",
fallbacks: ["anthropic/claude-sonnet-4-5", "google/gemini-2.5-flash"],
},
},
},
}),
provider: "anthropic",
model: "claude-sonnet-4-20250514",
calls: [
["anthropic", "claude-sonnet-4-20250514"],
["anthropic", "claude-sonnet-4-5"],
],
},
{
name: "same provider model version difference",
cfg: makeCfg({
agents: {
defaults: {
model: {
primary: "anthropic/claude-opus-4-6",
fallbacks: ["groq/llama-3.3-70b-versatile"],
},
},
},
}),
provider: "anthropic",
model: "claude-opus-4-5",
calls: [
["anthropic", "claude-opus-4-5"],
["groq", "llama-3.3-70b-versatile"],
],
},
{
name: "different provider skips configured fallbacks",
cfg: makeCfg({
agents: {
defaults: {
model: {
primary: "anthropic/claude-opus-4-6",
fallbacks: [],
},
},
},
}),
provider: "openai",
model: "gpt-4.1-mini",
calls: [
["openai", "gpt-4.1-mini"],
["anthropic", "claude-opus-4-6"],
],
},
{
name: "exact primary uses fallbacks",
cfg: makeCfg({
agents: {
defaults: {
model: {
primary: "anthropic/claude-opus-4-6",
fallbacks: ["groq/llama-3.3-70b-versatile"],
},
},
},
}),
provider: "anthropic",
model: "claude-opus-4-6",
calls: [
["anthropic", "claude-opus-4-6"],
["groq", "llama-3.3-70b-versatile"],
],
},
] satisfies Array<{
name: string;
cfg: OpenClawConfig;
provider: string;
model: string;
calls: Array<[string, string]>;
}>;
for (const testCase of cases) {
const candidates = __testing.resolveFallbackCandidates({
cfg: testCase.cfg,
provider: testCase.provider,
model: testCase.model,
});
expect(candidates.slice(0, testCase.calls.length), testCase.name).toEqual(
testCase.calls.map(([provider, model]) => ({ provider, model })),
);
}
});
});
describe("fallback behavior with provider cooldowns", () => {
async function makeAuthStoreWithCooldown(
provider: string,
reason: "rate_limit" | "overloaded" | "timeout" | "auth" | "billing",
): Promise<{ dir: string }> {
const tmpDir = await makeAuthTempDir();
const now = Date.now();
const store: AuthProfileStore = {
version: AUTH_STORE_VERSION,
profiles: {
[`${provider}:default`]: { type: "api_key", provider, key: "test-key" },
},
usageStats: {
[`${provider}:default`]:
reason === "rate_limit" || reason === "overloaded" || reason === "timeout"
? {
cooldownUntil: now + 300000,
failureCounts: { [reason]: 1 },
}
: {
disabledUntil: now + 300000,
disabledReason: reason,
},
},
};
setAuthRuntimeStore(tmpDir, store);
return { dir: tmpDir };
}
it("attempts same-provider fallbacks during transient cooldowns", async () => {
const { dir } = await makeAuthStoreWithCooldown("anthropic", "timeout");
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "anthropic/claude-opus-4-6",
fallbacks: ["anthropic/claude-sonnet-4-5", "groq/llama-3.3-70b-versatile"],
},
},
},
});
const run = vi.fn().mockResolvedValueOnce("sonnet success");
const result = await runWithModelFallback({
cfg,
provider: "anthropic",
model: "claude-opus-4-6",
run,
agentDir: dir,
});
expect(result.result).toBe("sonnet success");
expect(run).toHaveBeenCalledTimes(1);
expect(run).toHaveBeenNthCalledWith(1, "anthropic", "claude-sonnet-4-5", {
allowTransientCooldownProbe: true,
});
});
it("keeps alias-resolved primary models subject to transient cooldowns", async () => {
const { dir } = await makeAuthStoreWithCooldown("anthropic", "rate_limit");
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "anthropic/claude-sonnet-4-6",
fallbacks: ["anthropic/claude-haiku-3-5", "groq/llama-3.3-70b-versatile"],
},
models: {
"anthropic/claude-sonnet-4-6": { alias: "sonnet" },
},
},
},
});
const run = vi.fn().mockResolvedValueOnce("haiku success");
const result = await runWithModelFallback({
cfg,
provider: "anthropic",
model: "sonnet",
run,
agentDir: dir,
});
expect(result.result).toBe("haiku success");
expect(run).toHaveBeenCalledTimes(1);
expect(run).toHaveBeenNthCalledWith(1, "anthropic", "claude-haiku-3-5", {
allowTransientCooldownProbe: true,
});
});
it("skips same-provider models on persistent auth cooldowns", async () => {
const { dir } = await makeAuthStoreWithCooldown("anthropic", "auth");
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "anthropic/claude-opus-4-6",
fallbacks: ["anthropic/claude-sonnet-4-5", "groq/llama-3.3-70b-versatile"],
},
},
},
});
const run = vi.fn().mockResolvedValueOnce("groq success");
const result = await runWithModelFallback({
cfg,
provider: "anthropic",
model: "claude-opus-4-6",
run,
agentDir: dir,
});
expect(result.result).toBe("groq success");
expect(run).toHaveBeenCalledTimes(1);
expect(run).toHaveBeenNthCalledWith(1, "groq", "llama-3.3-70b-versatile");
});
it("tries cross-provider fallbacks when same provider has rate limit", async () => {
const tmpDir = await makeAuthTempDir();
const store: AuthProfileStore = {
version: AUTH_STORE_VERSION,
profiles: {
"anthropic:default": { type: "api_key", provider: "anthropic", key: "test-key" },
"groq:default": { type: "api_key", provider: "groq", key: "test-key" },
},
usageStats: {
"anthropic:default": {
cooldownUntil: Date.now() + 300000,
failureCounts: { rate_limit: 2 },
},
},
};
setAuthRuntimeStore(tmpDir, store);
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "anthropic/claude-opus-4-6",
fallbacks: ["anthropic/claude-sonnet-4-5", "groq/llama-3.3-70b-versatile"],
},
},
},
});
const run = vi
.fn()
.mockRejectedValueOnce(new Error("Still rate limited"))
.mockResolvedValueOnce("groq success");
const result = await runWithModelFallback({
cfg,
provider: "anthropic",
model: "claude-opus-4-6",
run,
agentDir: tmpDir,
});
expect(result.result).toBe("groq success");
expect(run).toHaveBeenCalledTimes(2);
expect(run).toHaveBeenNthCalledWith(1, "anthropic", "claude-sonnet-4-5", {
allowTransientCooldownProbe: true,
});
expect(run).toHaveBeenNthCalledWith(2, "groq", "llama-3.3-70b-versatile");
});
it("limits cooldown probes to one per provider before moving to cross-provider fallback", async () => {
const { dir } = await makeAuthStoreWithCooldown("anthropic", "rate_limit");
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "anthropic/claude-opus-4-6",
fallbacks: [
"anthropic/claude-sonnet-4-5",
"anthropic/claude-haiku-3-5",
"groq/llama-3.3-70b-versatile",
],
},
},
},
});
const run = vi
.fn()
.mockRejectedValueOnce(new Error("Still rate limited"))
.mockResolvedValueOnce("groq success");
const result = await runWithModelFallback({
cfg,
provider: "anthropic",
model: "claude-opus-4-6",
run,
agentDir: dir,
});
expect(result.result).toBe("groq success");
expect(run).toHaveBeenCalledTimes(2);
expect(run).toHaveBeenNthCalledWith(1, "anthropic", "claude-sonnet-4-5", {
allowTransientCooldownProbe: true,
});
expect(run).toHaveBeenNthCalledWith(2, "groq", "llama-3.3-70b-versatile");
});
it("does not consume transient probe slot when first same-provider probe fails with model_not_found", async () => {
const { dir } = await makeAuthStoreWithCooldown("anthropic", "rate_limit");
const cfg = makeCfg({
agents: {
defaults: {
model: {
primary: "anthropic/claude-opus-4-6",
fallbacks: [
"anthropic/claude-sonnet-4-5",
"anthropic/claude-haiku-3-5",
"groq/llama-3.3-70b-versatile",
],
},
},
},
});
const run = vi
.fn()
.mockRejectedValueOnce(new Error("Model not found: anthropic/claude-sonnet-4-5"))
.mockResolvedValueOnce("haiku success");
const result = await runWithModelFallback({
cfg,
provider: "anthropic",
model: "claude-opus-4-6",
run,
agentDir: dir,
});
expect(result.result).toBe("haiku success");
expect(run).toHaveBeenCalledTimes(2);
expect(run).toHaveBeenNthCalledWith(1, "anthropic", "claude-sonnet-4-5", {
allowTransientCooldownProbe: true,
});
expect(run).toHaveBeenNthCalledWith(2, "anthropic", "claude-haiku-3-5", {
allowTransientCooldownProbe: true,
});
});
});
});
describe("runWithImageModelFallback", () => {
it("resolves image-model override providers", async () => {
const cases = [
{
name: "bare override inherits configured provider",
cfg: makeCfg({
agents: {
defaults: {
imageModel: {
primary: "openai-codex/gpt-5.4",
fallbacks: ["openai-codex/gpt-5.4-mini"],
},
},
},
}),
modelOverride: "gpt-5.4-mini",
expected: [["openai-codex", "gpt-5.4-mini"]],
},
{
name: "qualified override keeps provider",
cfg: makeCfg({
agents: {
defaults: {
imageModel: {
primary: "openai-codex/gpt-5.4",
},
},
},
}),
modelOverride: "google/gemini-3-pro-image",
expected: [["google", "gemini-3-pro-image"]],
},
] satisfies Array<{
name: string;
cfg: OpenClawConfig;
modelOverride: string;
expected: Array<[string, string]>;
}>;
for (const testCase of cases) {
await runModelFallbackCase(testCase.name, async () => {
const run = vi.fn().mockResolvedValueOnce("ok");
const result = await runWithImageModelFallback({
cfg: testCase.cfg,
modelOverride: testCase.modelOverride,
run,
});
expect(result.result).toBe("ok");
expect(run.mock.calls).toEqual(testCase.expected);
});
}
});
it("keeps explicit image fallbacks reachable when models allowlist is present", async () => {
const cfg = makeCfg({
agents: {
defaults: {
imageModel: {
primary: "openai/gpt-image-1",
fallbacks: ["google/gemini-2.5-flash-image-preview"],
},
models: {
"openai/gpt-image-1": {},
},
},
},
});
const run = vi
.fn()
.mockRejectedValueOnce(new Error("rate limited"))
.mockResolvedValueOnce("ok");
const result = await runWithImageModelFallback({
cfg,
run,
});
expect(result.result).toBe("ok");
expect(run.mock.calls).toEqual([
["openai", "gpt-image-1"],
["google", "gemini-2.5-flash-image-preview"],
]);
});
});