mirror of
https://github.com/openclaw/openclaw.git
synced 2026-04-03 13:22:14 +00:00
* fix(providers): centralize media request shaping * style(providers): normalize shared request imports * fix(changelog): add media request shaping entry * fix(google): preserve private network guard
85 lines
2.3 KiB
TypeScript
85 lines
2.3 KiB
TypeScript
import type {
|
|
AudioTranscriptionRequest,
|
|
AudioTranscriptionResult,
|
|
} from "openclaw/plugin-sdk/media-understanding";
|
|
import {
|
|
assertOkOrThrowHttpError,
|
|
postTranscriptionRequest,
|
|
resolveProviderHttpRequestConfig,
|
|
requireTranscriptionText,
|
|
} from "openclaw/plugin-sdk/provider-http";
|
|
|
|
export const DEFAULT_DEEPGRAM_AUDIO_BASE_URL = "https://api.deepgram.com/v1";
|
|
export const DEFAULT_DEEPGRAM_AUDIO_MODEL = "nova-3";
|
|
|
|
function resolveModel(model?: string): string {
|
|
const trimmed = model?.trim();
|
|
return trimmed || DEFAULT_DEEPGRAM_AUDIO_MODEL;
|
|
}
|
|
|
|
type DeepgramTranscriptResponse = {
|
|
results?: {
|
|
channels?: Array<{
|
|
alternatives?: Array<{
|
|
transcript?: string;
|
|
}>;
|
|
}>;
|
|
};
|
|
};
|
|
|
|
export async function transcribeDeepgramAudio(
|
|
params: AudioTranscriptionRequest,
|
|
): Promise<AudioTranscriptionResult> {
|
|
const fetchFn = params.fetchFn ?? fetch;
|
|
const model = resolveModel(params.model);
|
|
const { baseUrl, allowPrivateNetwork, headers } = resolveProviderHttpRequestConfig({
|
|
baseUrl: params.baseUrl,
|
|
defaultBaseUrl: DEFAULT_DEEPGRAM_AUDIO_BASE_URL,
|
|
headers: params.headers,
|
|
defaultHeaders: {
|
|
authorization: `Token ${params.apiKey}`,
|
|
"content-type": params.mime ?? "application/octet-stream",
|
|
},
|
|
provider: "deepgram",
|
|
capability: "audio",
|
|
transport: "media-understanding",
|
|
});
|
|
|
|
const url = new URL(`${baseUrl}/listen`);
|
|
url.searchParams.set("model", model);
|
|
if (params.language?.trim()) {
|
|
url.searchParams.set("language", params.language.trim());
|
|
}
|
|
if (params.query) {
|
|
for (const [key, value] of Object.entries(params.query)) {
|
|
if (value === undefined) {
|
|
continue;
|
|
}
|
|
url.searchParams.set(key, String(value));
|
|
}
|
|
}
|
|
|
|
const body = new Uint8Array(params.buffer);
|
|
const { response: res, release } = await postTranscriptionRequest({
|
|
url: url.toString(),
|
|
headers,
|
|
body,
|
|
timeoutMs: params.timeoutMs,
|
|
fetchFn,
|
|
allowPrivateNetwork,
|
|
});
|
|
|
|
try {
|
|
await assertOkOrThrowHttpError(res, "Audio transcription failed");
|
|
|
|
const payload = (await res.json()) as DeepgramTranscriptResponse;
|
|
const transcript = requireTranscriptionText(
|
|
payload.results?.channels?.[0]?.alternatives?.[0]?.transcript,
|
|
"Audio transcription response missing transcript",
|
|
);
|
|
return { text: transcript, model };
|
|
} finally {
|
|
await release();
|
|
}
|
|
}
|