mirror of
https://github.com/openclaw/openclaw.git
synced 2026-03-13 19:10:39 +00:00
repairToolUseResultPairing was gated behind !isOpenAi, skipping orphaned
tool_result cleanup for OpenAI providers. When limitHistoryTurns truncated
conversation history, tool_result messages whose matching tool_call was
before the truncation point survived and were sent as function_call_output
items with stale call_id references. OpenAI rejects these with:
"No tool call found for function call output with call_id ..."
Enable the repair universally — all providers need it after truncation.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
(cherry picked from commit 97b065aa6e)
133 lines
4.8 KiB
TypeScript
133 lines
4.8 KiB
TypeScript
import { normalizeProviderId } from "./model-selection.js";
|
|
import { isGoogleModelApi } from "./pi-embedded-helpers/google.js";
|
|
import type { ToolCallIdMode } from "./tool-call-id.js";
|
|
|
|
export type TranscriptSanitizeMode = "full" | "images-only";
|
|
|
|
export type TranscriptPolicy = {
|
|
sanitizeMode: TranscriptSanitizeMode;
|
|
sanitizeToolCallIds: boolean;
|
|
toolCallIdMode?: ToolCallIdMode;
|
|
repairToolUseResultPairing: boolean;
|
|
preserveSignatures: boolean;
|
|
sanitizeThoughtSignatures?: {
|
|
allowBase64Only?: boolean;
|
|
includeCamelCase?: boolean;
|
|
};
|
|
sanitizeThinkingSignatures: boolean;
|
|
dropThinkingBlocks: boolean;
|
|
applyGoogleTurnOrdering: boolean;
|
|
validateGeminiTurns: boolean;
|
|
validateAnthropicTurns: boolean;
|
|
allowSyntheticToolResults: boolean;
|
|
};
|
|
|
|
const MISTRAL_MODEL_HINTS = [
|
|
"mistral",
|
|
"mixtral",
|
|
"codestral",
|
|
"pixtral",
|
|
"devstral",
|
|
"ministral",
|
|
"mistralai",
|
|
];
|
|
const OPENAI_MODEL_APIS = new Set([
|
|
"openai",
|
|
"openai-completions",
|
|
"openai-responses",
|
|
"openai-codex-responses",
|
|
]);
|
|
const OPENAI_PROVIDERS = new Set(["openai", "openai-codex"]);
|
|
const OPENAI_COMPAT_TURN_MERGE_EXCLUDED_PROVIDERS = new Set(["openrouter", "opencode"]);
|
|
|
|
function isOpenAiApi(modelApi?: string | null): boolean {
|
|
if (!modelApi) {
|
|
return false;
|
|
}
|
|
return OPENAI_MODEL_APIS.has(modelApi);
|
|
}
|
|
|
|
function isOpenAiProvider(provider?: string | null): boolean {
|
|
if (!provider) {
|
|
return false;
|
|
}
|
|
return OPENAI_PROVIDERS.has(normalizeProviderId(provider));
|
|
}
|
|
|
|
function isAnthropicApi(modelApi?: string | null, provider?: string | null): boolean {
|
|
if (modelApi === "anthropic-messages" || modelApi === "bedrock-converse-stream") {
|
|
return true;
|
|
}
|
|
const normalized = normalizeProviderId(provider ?? "");
|
|
// MiniMax now uses openai-completions API, not anthropic-messages
|
|
return normalized === "anthropic" || normalized === "amazon-bedrock";
|
|
}
|
|
|
|
function isMistralModel(params: { provider?: string | null; modelId?: string | null }): boolean {
|
|
const provider = normalizeProviderId(params.provider ?? "");
|
|
if (provider === "mistral") {
|
|
return true;
|
|
}
|
|
const modelId = (params.modelId ?? "").toLowerCase();
|
|
if (!modelId) {
|
|
return false;
|
|
}
|
|
return MISTRAL_MODEL_HINTS.some((hint) => modelId.includes(hint));
|
|
}
|
|
|
|
export function resolveTranscriptPolicy(params: {
|
|
modelApi?: string | null;
|
|
provider?: string | null;
|
|
modelId?: string | null;
|
|
}): TranscriptPolicy {
|
|
const provider = normalizeProviderId(params.provider ?? "");
|
|
const modelId = params.modelId ?? "";
|
|
const isGoogle = isGoogleModelApi(params.modelApi);
|
|
const isAnthropic = isAnthropicApi(params.modelApi, provider);
|
|
const isOpenAi = isOpenAiProvider(provider) || (!provider && isOpenAiApi(params.modelApi));
|
|
const isStrictOpenAiCompatible =
|
|
params.modelApi === "openai-completions" &&
|
|
!isOpenAi &&
|
|
!OPENAI_COMPAT_TURN_MERGE_EXCLUDED_PROVIDERS.has(provider);
|
|
const isMistral = isMistralModel({ provider, modelId });
|
|
const isOpenRouterGemini =
|
|
(provider === "openrouter" || provider === "opencode" || provider === "kilocode") &&
|
|
modelId.toLowerCase().includes("gemini");
|
|
const isCopilotClaude = provider === "github-copilot" && modelId.toLowerCase().includes("claude");
|
|
|
|
// GitHub Copilot's Claude endpoints can reject persisted `thinking` blocks with
|
|
// non-binary/non-base64 signatures (e.g. thinkingSignature: "reasoning_text").
|
|
// Drop these blocks at send-time to keep sessions usable.
|
|
const dropThinkingBlocks = isCopilotClaude;
|
|
|
|
const needsNonImageSanitize = isGoogle || isAnthropic || isMistral || isOpenRouterGemini;
|
|
|
|
const sanitizeToolCallIds = isGoogle || isMistral || isAnthropic;
|
|
const toolCallIdMode: ToolCallIdMode | undefined = isMistral
|
|
? "strict9"
|
|
: sanitizeToolCallIds
|
|
? "strict"
|
|
: undefined;
|
|
// All providers need orphaned tool_result repair after history truncation.
|
|
// OpenAI rejects function_call_output items whose call_id has no matching
|
|
// function_call in the conversation, so the repair must run universally.
|
|
const repairToolUseResultPairing = true;
|
|
const sanitizeThoughtSignatures =
|
|
isOpenRouterGemini || isGoogle ? { allowBase64Only: true, includeCamelCase: true } : undefined;
|
|
|
|
return {
|
|
sanitizeMode: isOpenAi ? "images-only" : needsNonImageSanitize ? "full" : "images-only",
|
|
sanitizeToolCallIds: !isOpenAi && sanitizeToolCallIds,
|
|
toolCallIdMode,
|
|
repairToolUseResultPairing,
|
|
preserveSignatures: false,
|
|
sanitizeThoughtSignatures: isOpenAi ? undefined : sanitizeThoughtSignatures,
|
|
sanitizeThinkingSignatures: false,
|
|
dropThinkingBlocks,
|
|
applyGoogleTurnOrdering: !isOpenAi && isGoogle,
|
|
validateGeminiTurns: !isOpenAi && isGoogle,
|
|
validateAnthropicTurns: !isOpenAi && (isAnthropic || isStrictOpenAiCompatible),
|
|
allowSyntheticToolResults: !isOpenAi && (isGoogle || isAnthropic),
|
|
};
|
|
}
|