import { afterEach, describe, expect, it, vi } from "vitest"; import { buildOpenAIVideoGenerationProvider } from "./video-generation-provider.js"; const { resolveApiKeyForProviderMock, postJsonRequestMock, fetchWithTimeoutMock, assertOkOrThrowHttpErrorMock, resolveProviderHttpRequestConfigMock, } = vi.hoisted(() => ({ resolveApiKeyForProviderMock: vi.fn(async () => ({ apiKey: "openai-key" })), postJsonRequestMock: vi.fn(), fetchWithTimeoutMock: vi.fn(), assertOkOrThrowHttpErrorMock: vi.fn(async () => {}), resolveProviderHttpRequestConfigMock: vi.fn((params) => ({ baseUrl: params.baseUrl ?? params.defaultBaseUrl, allowPrivateNetwork: false, headers: new Headers(params.defaultHeaders), dispatcherPolicy: undefined, })), })); vi.mock("openclaw/plugin-sdk/provider-auth-runtime", () => ({ resolveApiKeyForProvider: resolveApiKeyForProviderMock, })); vi.mock("openclaw/plugin-sdk/provider-http", () => ({ assertOkOrThrowHttpError: assertOkOrThrowHttpErrorMock, fetchWithTimeout: fetchWithTimeoutMock, postJsonRequest: postJsonRequestMock, resolveProviderHttpRequestConfig: resolveProviderHttpRequestConfigMock, })); describe("openai video generation provider", () => { afterEach(() => { resolveApiKeyForProviderMock.mockClear(); postJsonRequestMock.mockReset(); fetchWithTimeoutMock.mockReset(); assertOkOrThrowHttpErrorMock.mockClear(); resolveProviderHttpRequestConfigMock.mockClear(); }); it("uses JSON for text-only Sora requests", async () => { postJsonRequestMock.mockResolvedValue({ response: { json: async () => ({ id: "vid_123", model: "sora-2", status: "queued", }), }, release: vi.fn(async () => {}), }); fetchWithTimeoutMock .mockResolvedValueOnce({ json: async () => ({ id: "vid_123", model: "sora-2", status: "completed", seconds: "4", size: "720x1280", }), }) .mockResolvedValueOnce({ headers: new Headers({ "content-type": "video/mp4" }), arrayBuffer: async () => Buffer.from("mp4-bytes"), }); const provider = buildOpenAIVideoGenerationProvider(); const result = await provider.generateVideo({ provider: "openai", model: "sora-2", prompt: "A paper airplane gliding through golden hour light", cfg: {}, durationSeconds: 4, }); expect(postJsonRequestMock).toHaveBeenCalledWith( expect.objectContaining({ url: "https://api.openai.com/v1/videos", }), ); expect(fetchWithTimeoutMock).toHaveBeenNthCalledWith( 1, "https://api.openai.com/v1/videos/vid_123", expect.objectContaining({ method: "GET" }), 120000, fetch, ); expect(result.videos).toHaveLength(1); expect(result.videos[0]?.mimeType).toBe("video/mp4"); expect(result.metadata).toEqual( expect.objectContaining({ videoId: "vid_123", status: "completed", }), ); }); it("uses JSON input_reference.image_url for image-to-video requests", async () => { postJsonRequestMock.mockResolvedValue({ response: { json: async () => ({ id: "vid_456", model: "sora-2", status: "queued", }), }, release: vi.fn(async () => {}), }); fetchWithTimeoutMock .mockResolvedValueOnce({ json: async () => ({ id: "vid_456", model: "sora-2", status: "completed", }), }) .mockResolvedValueOnce({ headers: new Headers({ "content-type": "video/mp4" }), arrayBuffer: async () => Buffer.from("mp4-bytes"), }); const provider = buildOpenAIVideoGenerationProvider(); await provider.generateVideo({ provider: "openai", model: "sora-2", prompt: "Animate this frame", cfg: {}, inputImages: [{ buffer: Buffer.from("png-bytes"), mimeType: "image/png" }], }); expect(postJsonRequestMock).toHaveBeenCalledWith( expect.objectContaining({ url: "https://api.openai.com/v1/videos", body: expect.objectContaining({ input_reference: { image_url: "data:image/png;base64,cG5nLWJ5dGVz", }, }), }), ); expect(fetchWithTimeoutMock).toHaveBeenNthCalledWith( 1, "https://api.openai.com/v1/videos/vid_456", expect.objectContaining({ method: "GET", }), 120000, fetch, ); }); it("honors configured baseUrl for video requests", async () => { postJsonRequestMock.mockResolvedValue({ response: { json: async () => ({ id: "vid_local", model: "sora-2", status: "queued", }), }, release: vi.fn(async () => {}), }); fetchWithTimeoutMock .mockResolvedValueOnce({ json: async () => ({ id: "vid_local", model: "sora-2", status: "completed", }), }) .mockResolvedValueOnce({ headers: new Headers({ "content-type": "video/mp4" }), arrayBuffer: async () => Buffer.from("mp4-bytes"), }); const provider = buildOpenAIVideoGenerationProvider(); await provider.generateVideo({ provider: "openai", model: "sora-2", prompt: "Render via local relay", cfg: { models: { providers: { openai: { baseUrl: "http://127.0.0.1:44080/v1", models: [], }, }, }, }, }); expect(resolveProviderHttpRequestConfigMock).toHaveBeenCalledWith( expect.objectContaining({ baseUrl: "http://127.0.0.1:44080/v1", }), ); expect(postJsonRequestMock).toHaveBeenCalledWith( expect.objectContaining({ url: "http://127.0.0.1:44080/v1/videos", allowPrivateNetwork: false, }), ); }); it("uses multipart input_reference for video-to-video uploads", async () => { fetchWithTimeoutMock .mockResolvedValueOnce({ ok: true, json: async () => ({ id: "vid_789", model: "sora-2", status: "queued", }), }) .mockResolvedValueOnce({ json: async () => ({ id: "vid_789", model: "sora-2", status: "completed", }), }) .mockResolvedValueOnce({ headers: new Headers({ "content-type": "video/mp4" }), arrayBuffer: async () => Buffer.from("mp4-bytes"), }); const provider = buildOpenAIVideoGenerationProvider(); await provider.generateVideo({ provider: "openai", model: "sora-2", prompt: "Remix this clip", cfg: {}, inputVideos: [{ buffer: Buffer.from("mp4-bytes"), mimeType: "video/mp4" }], }); expect(postJsonRequestMock).not.toHaveBeenCalled(); expect(fetchWithTimeoutMock).toHaveBeenNthCalledWith( 1, "https://api.openai.com/v1/videos", expect.objectContaining({ method: "POST", body: expect.any(FormData), }), 120000, fetch, ); }); it("rejects multiple reference assets", async () => { const provider = buildOpenAIVideoGenerationProvider(); await expect( provider.generateVideo({ provider: "openai", model: "sora-2", prompt: "Animate these", cfg: {}, inputImages: [{ buffer: Buffer.from("a"), mimeType: "image/png" }], inputVideos: [{ buffer: Buffer.from("b"), mimeType: "video/mp4" }], }), ).rejects.toThrow("OpenAI video generation supports at most one reference image or video."); }); });