mirror of
https://github.com/openclaw/openclaw.git
synced 2026-03-12 07:20:45 +00:00
Merged via squash.
Prepared head SHA: a994c07190
Co-authored-by: gumadeiras <5599352+gumadeiras@users.noreply.github.com>
Co-authored-by: gumadeiras <5599352+gumadeiras@users.noreply.github.com>
Reviewed-by: @gumadeiras
325 lines
11 KiB
TypeScript
325 lines
11 KiB
TypeScript
import fsSync from "node:fs";
|
|
import type { Llama, LlamaEmbeddingContext, LlamaModel } from "node-llama-cpp";
|
|
import type { OpenClawConfig } from "../config/config.js";
|
|
import type { SecretInput } from "../config/types.secrets.js";
|
|
import { formatErrorMessage } from "../infra/errors.js";
|
|
import { resolveUserPath } from "../utils.js";
|
|
import type { EmbeddingInput } from "./embedding-inputs.js";
|
|
import { sanitizeAndNormalizeEmbedding } from "./embedding-vectors.js";
|
|
import {
|
|
createGeminiEmbeddingProvider,
|
|
type GeminiEmbeddingClient,
|
|
type GeminiTaskType,
|
|
} from "./embeddings-gemini.js";
|
|
import {
|
|
createMistralEmbeddingProvider,
|
|
type MistralEmbeddingClient,
|
|
} from "./embeddings-mistral.js";
|
|
import { createOllamaEmbeddingProvider, type OllamaEmbeddingClient } from "./embeddings-ollama.js";
|
|
import { createOpenAiEmbeddingProvider, type OpenAiEmbeddingClient } from "./embeddings-openai.js";
|
|
import { createVoyageEmbeddingProvider, type VoyageEmbeddingClient } from "./embeddings-voyage.js";
|
|
import { importNodeLlamaCpp } from "./node-llama.js";
|
|
|
|
export type { GeminiEmbeddingClient } from "./embeddings-gemini.js";
|
|
export type { MistralEmbeddingClient } from "./embeddings-mistral.js";
|
|
export type { OpenAiEmbeddingClient } from "./embeddings-openai.js";
|
|
export type { VoyageEmbeddingClient } from "./embeddings-voyage.js";
|
|
export type { OllamaEmbeddingClient } from "./embeddings-ollama.js";
|
|
|
|
export type EmbeddingProvider = {
|
|
id: string;
|
|
model: string;
|
|
maxInputTokens?: number;
|
|
embedQuery: (text: string) => Promise<number[]>;
|
|
embedBatch: (texts: string[]) => Promise<number[][]>;
|
|
embedBatchInputs?: (inputs: EmbeddingInput[]) => Promise<number[][]>;
|
|
};
|
|
|
|
export type EmbeddingProviderId = "openai" | "local" | "gemini" | "voyage" | "mistral" | "ollama";
|
|
export type EmbeddingProviderRequest = EmbeddingProviderId | "auto";
|
|
export type EmbeddingProviderFallback = EmbeddingProviderId | "none";
|
|
|
|
// Remote providers considered for auto-selection when provider === "auto".
|
|
// Ollama is intentionally excluded here so that "auto" mode does not
|
|
// implicitly assume a local Ollama instance is available.
|
|
const REMOTE_EMBEDDING_PROVIDER_IDS = ["openai", "gemini", "voyage", "mistral"] as const;
|
|
|
|
export type EmbeddingProviderResult = {
|
|
provider: EmbeddingProvider | null;
|
|
requestedProvider: EmbeddingProviderRequest;
|
|
fallbackFrom?: EmbeddingProviderId;
|
|
fallbackReason?: string;
|
|
providerUnavailableReason?: string;
|
|
openAi?: OpenAiEmbeddingClient;
|
|
gemini?: GeminiEmbeddingClient;
|
|
voyage?: VoyageEmbeddingClient;
|
|
mistral?: MistralEmbeddingClient;
|
|
ollama?: OllamaEmbeddingClient;
|
|
};
|
|
|
|
export type EmbeddingProviderOptions = {
|
|
config: OpenClawConfig;
|
|
agentDir?: string;
|
|
provider: EmbeddingProviderRequest;
|
|
remote?: {
|
|
baseUrl?: string;
|
|
apiKey?: SecretInput;
|
|
headers?: Record<string, string>;
|
|
};
|
|
model: string;
|
|
fallback: EmbeddingProviderFallback;
|
|
local?: {
|
|
modelPath?: string;
|
|
modelCacheDir?: string;
|
|
};
|
|
/** Gemini embedding-2: output vector dimensions (768, 1536, or 3072). */
|
|
outputDimensionality?: number;
|
|
/** Gemini: override the default task type sent with embedding requests. */
|
|
taskType?: GeminiTaskType;
|
|
};
|
|
|
|
export const DEFAULT_LOCAL_MODEL =
|
|
"hf:ggml-org/embeddinggemma-300m-qat-q8_0-GGUF/embeddinggemma-300m-qat-Q8_0.gguf";
|
|
|
|
function canAutoSelectLocal(options: EmbeddingProviderOptions): boolean {
|
|
const modelPath = options.local?.modelPath?.trim();
|
|
if (!modelPath) {
|
|
return false;
|
|
}
|
|
if (/^(hf:|https?:)/i.test(modelPath)) {
|
|
return false;
|
|
}
|
|
const resolved = resolveUserPath(modelPath);
|
|
try {
|
|
return fsSync.statSync(resolved).isFile();
|
|
} catch {
|
|
return false;
|
|
}
|
|
}
|
|
|
|
function isMissingApiKeyError(err: unknown): boolean {
|
|
const message = formatErrorMessage(err);
|
|
return message.includes("No API key found for provider");
|
|
}
|
|
|
|
async function createLocalEmbeddingProvider(
|
|
options: EmbeddingProviderOptions,
|
|
): Promise<EmbeddingProvider> {
|
|
const modelPath = options.local?.modelPath?.trim() || DEFAULT_LOCAL_MODEL;
|
|
const modelCacheDir = options.local?.modelCacheDir?.trim();
|
|
|
|
// Lazy-load node-llama-cpp to keep startup light unless local is enabled.
|
|
const { getLlama, resolveModelFile, LlamaLogLevel } = await importNodeLlamaCpp();
|
|
|
|
let llama: Llama | null = null;
|
|
let embeddingModel: LlamaModel | null = null;
|
|
let embeddingContext: LlamaEmbeddingContext | null = null;
|
|
let initPromise: Promise<LlamaEmbeddingContext> | null = null;
|
|
|
|
const ensureContext = async (): Promise<LlamaEmbeddingContext> => {
|
|
if (embeddingContext) {
|
|
return embeddingContext;
|
|
}
|
|
if (initPromise) {
|
|
return initPromise;
|
|
}
|
|
initPromise = (async () => {
|
|
try {
|
|
if (!llama) {
|
|
llama = await getLlama({ logLevel: LlamaLogLevel.error });
|
|
}
|
|
if (!embeddingModel) {
|
|
const resolved = await resolveModelFile(modelPath, modelCacheDir || undefined);
|
|
embeddingModel = await llama.loadModel({ modelPath: resolved });
|
|
}
|
|
if (!embeddingContext) {
|
|
embeddingContext = await embeddingModel.createEmbeddingContext();
|
|
}
|
|
return embeddingContext;
|
|
} catch (err) {
|
|
initPromise = null;
|
|
throw err;
|
|
}
|
|
})();
|
|
return initPromise;
|
|
};
|
|
|
|
return {
|
|
id: "local",
|
|
model: modelPath,
|
|
embedQuery: async (text) => {
|
|
const ctx = await ensureContext();
|
|
const embedding = await ctx.getEmbeddingFor(text);
|
|
return sanitizeAndNormalizeEmbedding(Array.from(embedding.vector));
|
|
},
|
|
embedBatch: async (texts) => {
|
|
const ctx = await ensureContext();
|
|
const embeddings = await Promise.all(
|
|
texts.map(async (text) => {
|
|
const embedding = await ctx.getEmbeddingFor(text);
|
|
return sanitizeAndNormalizeEmbedding(Array.from(embedding.vector));
|
|
}),
|
|
);
|
|
return embeddings;
|
|
},
|
|
};
|
|
}
|
|
|
|
export async function createEmbeddingProvider(
|
|
options: EmbeddingProviderOptions,
|
|
): Promise<EmbeddingProviderResult> {
|
|
const requestedProvider = options.provider;
|
|
const fallback = options.fallback;
|
|
|
|
const createProvider = async (id: EmbeddingProviderId) => {
|
|
if (id === "local") {
|
|
const provider = await createLocalEmbeddingProvider(options);
|
|
return { provider };
|
|
}
|
|
if (id === "ollama") {
|
|
const { provider, client } = await createOllamaEmbeddingProvider(options);
|
|
return { provider, ollama: client };
|
|
}
|
|
if (id === "gemini") {
|
|
const { provider, client } = await createGeminiEmbeddingProvider(options);
|
|
return { provider, gemini: client };
|
|
}
|
|
if (id === "voyage") {
|
|
const { provider, client } = await createVoyageEmbeddingProvider(options);
|
|
return { provider, voyage: client };
|
|
}
|
|
if (id === "mistral") {
|
|
const { provider, client } = await createMistralEmbeddingProvider(options);
|
|
return { provider, mistral: client };
|
|
}
|
|
const { provider, client } = await createOpenAiEmbeddingProvider(options);
|
|
return { provider, openAi: client };
|
|
};
|
|
|
|
const formatPrimaryError = (err: unknown, provider: EmbeddingProviderId) =>
|
|
provider === "local" ? formatLocalSetupError(err) : formatErrorMessage(err);
|
|
|
|
if (requestedProvider === "auto") {
|
|
const missingKeyErrors: string[] = [];
|
|
let localError: string | null = null;
|
|
|
|
if (canAutoSelectLocal(options)) {
|
|
try {
|
|
const local = await createProvider("local");
|
|
return { ...local, requestedProvider };
|
|
} catch (err) {
|
|
localError = formatLocalSetupError(err);
|
|
}
|
|
}
|
|
|
|
for (const provider of REMOTE_EMBEDDING_PROVIDER_IDS) {
|
|
try {
|
|
const result = await createProvider(provider);
|
|
return { ...result, requestedProvider };
|
|
} catch (err) {
|
|
const message = formatPrimaryError(err, provider);
|
|
if (isMissingApiKeyError(err)) {
|
|
missingKeyErrors.push(message);
|
|
continue;
|
|
}
|
|
// Non-auth errors (e.g., network) are still fatal
|
|
const wrapped = new Error(message) as Error & { cause?: unknown };
|
|
wrapped.cause = err;
|
|
throw wrapped;
|
|
}
|
|
}
|
|
|
|
// All providers failed due to missing API keys - return null provider for FTS-only mode
|
|
const details = [...missingKeyErrors, localError].filter(Boolean) as string[];
|
|
const reason = details.length > 0 ? details.join("\n\n") : "No embeddings provider available.";
|
|
return {
|
|
provider: null,
|
|
requestedProvider,
|
|
providerUnavailableReason: reason,
|
|
};
|
|
}
|
|
|
|
try {
|
|
const primary = await createProvider(requestedProvider);
|
|
return { ...primary, requestedProvider };
|
|
} catch (primaryErr) {
|
|
const reason = formatPrimaryError(primaryErr, requestedProvider);
|
|
if (fallback && fallback !== "none" && fallback !== requestedProvider) {
|
|
try {
|
|
const fallbackResult = await createProvider(fallback);
|
|
return {
|
|
...fallbackResult,
|
|
requestedProvider,
|
|
fallbackFrom: requestedProvider,
|
|
fallbackReason: reason,
|
|
};
|
|
} catch (fallbackErr) {
|
|
// Both primary and fallback failed - check if it's auth-related
|
|
const fallbackReason = formatErrorMessage(fallbackErr);
|
|
const combinedReason = `${reason}\n\nFallback to ${fallback} failed: ${fallbackReason}`;
|
|
if (isMissingApiKeyError(primaryErr) && isMissingApiKeyError(fallbackErr)) {
|
|
// Both failed due to missing API keys - return null for FTS-only mode
|
|
return {
|
|
provider: null,
|
|
requestedProvider,
|
|
fallbackFrom: requestedProvider,
|
|
fallbackReason: reason,
|
|
providerUnavailableReason: combinedReason,
|
|
};
|
|
}
|
|
// Non-auth errors are still fatal
|
|
const wrapped = new Error(combinedReason) as Error & { cause?: unknown };
|
|
wrapped.cause = fallbackErr;
|
|
throw wrapped;
|
|
}
|
|
}
|
|
// No fallback configured - check if we should degrade to FTS-only
|
|
if (isMissingApiKeyError(primaryErr)) {
|
|
return {
|
|
provider: null,
|
|
requestedProvider,
|
|
providerUnavailableReason: reason,
|
|
};
|
|
}
|
|
const wrapped = new Error(reason) as Error & { cause?: unknown };
|
|
wrapped.cause = primaryErr;
|
|
throw wrapped;
|
|
}
|
|
}
|
|
|
|
function isNodeLlamaCppMissing(err: unknown): boolean {
|
|
if (!(err instanceof Error)) {
|
|
return false;
|
|
}
|
|
const code = (err as Error & { code?: unknown }).code;
|
|
if (code === "ERR_MODULE_NOT_FOUND") {
|
|
return err.message.includes("node-llama-cpp");
|
|
}
|
|
return false;
|
|
}
|
|
|
|
function formatLocalSetupError(err: unknown): string {
|
|
const detail = formatErrorMessage(err);
|
|
const missing = isNodeLlamaCppMissing(err);
|
|
return [
|
|
"Local embeddings unavailable.",
|
|
missing
|
|
? "Reason: optional dependency node-llama-cpp is missing (or failed to install)."
|
|
: detail
|
|
? `Reason: ${detail}`
|
|
: undefined,
|
|
missing && detail ? `Detail: ${detail}` : null,
|
|
"To enable local embeddings:",
|
|
"1) Use Node 22 LTS (recommended for installs/updates)",
|
|
missing
|
|
? "2) Reinstall OpenClaw (this should install node-llama-cpp): npm i -g openclaw@latest"
|
|
: null,
|
|
"3) If you use pnpm: pnpm approve-builds (select node-llama-cpp), then pnpm rebuild node-llama-cpp",
|
|
...REMOTE_EMBEDDING_PROVIDER_IDS.map(
|
|
(provider) => `Or set agents.defaults.memorySearch.provider = "${provider}" (remote).`,
|
|
),
|
|
]
|
|
.filter(Boolean)
|
|
.join("\n");
|
|
}
|