import type { ProviderResolveDynamicModelContext, ProviderRuntimeModel, } from "openclaw/plugin-sdk/core"; import { normalizeModelCompat } from "openclaw/plugin-sdk/provider-models"; export const PROVIDER_ID = "github-copilot"; const CODEX_GPT_53_MODEL_ID = "gpt-5.3-codex"; const CODEX_TEMPLATE_MODEL_IDS = ["gpt-5.2-codex"] as const; const DEFAULT_CONTEXT_WINDOW = 128_000; const DEFAULT_MAX_TOKENS = 8192; export function resolveCopilotForwardCompatModel( ctx: ProviderResolveDynamicModelContext, ): ProviderRuntimeModel | undefined { const trimmedModelId = ctx.modelId.trim(); if (!trimmedModelId) { return undefined; } // If the model is already in the registry, let the normal path handle it. const existing = ctx.modelRegistry.find(PROVIDER_ID, trimmedModelId.toLowerCase()); if (existing) { return undefined; } // For gpt-5.3-codex specifically, clone from the gpt-5.2-codex template // to preserve any special settings the registry has for codex models. if (trimmedModelId.toLowerCase() === CODEX_GPT_53_MODEL_ID) { for (const templateId of CODEX_TEMPLATE_MODEL_IDS) { const template = ctx.modelRegistry.find( PROVIDER_ID, templateId, ) as ProviderRuntimeModel | null; if (!template) { continue; } return normalizeModelCompat({ ...template, id: trimmedModelId, name: trimmedModelId, } as ProviderRuntimeModel); } // Template not found — fall through to synthetic catch-all below. } // Catch-all: create a synthetic model definition for any unknown model ID. // The Copilot API is OpenAI-compatible and will return its own error if the // model isn't available on the user's plan. This lets new models be used // by simply adding them to agents.defaults.models in openclaw.json — no // code change required. const lowerModelId = trimmedModelId.toLowerCase(); const reasoning = /^o[13](\b|$)/.test(lowerModelId); return normalizeModelCompat({ id: trimmedModelId, name: trimmedModelId, provider: PROVIDER_ID, api: "openai-responses", reasoning, // Optimistic: most Copilot models support images, and the API rejects // image payloads for text-only models rather than failing silently. input: ["text", "image"], cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 }, contextWindow: DEFAULT_CONTEXT_WINDOW, maxTokens: DEFAULT_MAX_TOKENS, } as ProviderRuntimeModel); }