Files
9router/open-sse/handlers/embeddingsCore.js
HXD.VN e1b836168a feat: add /v1/embeddings endpoint (OpenAI-compatible) (#146)
* feat: implement /v1/embeddings endpoint (#117)

Add OpenAI-compatible POST /v1/embeddings endpoint that routes through
the existing provider credential + fallback infrastructure.

Changes:
- open-sse/handlers/embeddingsCore.js: core handler (handleEmbeddingsCore)
  * Validates input (string or array), encoding_format
  * Builds provider-specific URL and headers for openai, openrouter,
    and openai-compatible providers
  * Handles 401/403 token refresh via executor.refreshCredentials
  * Returns normalized OpenAI-format response { object: 'list', data, model, usage }
- cloud/src/handlers/embeddings.js: cloud Worker handler (handleEmbeddings)
  * Auth + machineId resolution identical to handleChat
  * Provider credential fallback loop with rate-limit tracking
- cloud/src/index.js: wire new routes
  * POST /v1/embeddings  (new format — machineId from API key)
  * POST /{machineId}/v1/embeddings  (old format — machineId from URL)

* test: add unit tests for /v1/embeddings endpoint

- Setup vitest as test framework (tests/ directory)
- embeddingsCore.test.js (36 tests):
  - buildEmbeddingsBody: single string, array, encoding_format, default float
  - buildEmbeddingsUrl: openai, openrouter, openai-compatible-*, unsupported
  - buildEmbeddingsHeaders: per-provider headers, accessToken fallback
  - handleEmbeddingsCore: input validation, success path, provider errors,
    network errors, invalid JSON, token refresh 401 handling
- embeddings.cloud.test.js (23 tests):
  - CORS OPTIONS preflight
  - Auth: missing/invalid/old-format/wrong key → 401/400
  - Body validation: bad JSON, missing model, missing input, bad model → 400
  - Happy path: single string, array, delegation, CORS header, machineId override
  - Rate limiting: all-rate-limited → 429 + Retry-After, no credentials → 400
  - Error propagation: non-fallback errors, 429 exhausts accounts

Total: 59/59 tests passing
Framework: vitest v4.0.18, Node v22.22.0

* feat: add Next.js API route for /v1/embeddings endpoint

Wire the embeddings handler into Next.js App Router.

- src/app/api/v1/embeddings/route.js: Next.js API route (POST + OPTIONS)
- src/sse/handlers/embeddings.js: SSE-layer handler mirroring chat.js pattern

Uses handleEmbeddingsCore from open-sse/handlers/embeddingsCore.js with
the same auth, credential fallback, and token refresh logic as the chat
handler. Supports REQUIRE_API_KEY env var, provider fallback loop, and
consistent logging.
2026-02-18 13:24:02 +07:00

210 lines
6.8 KiB
JavaScript

import { getModelTargetFormat, PROVIDER_ID_TO_ALIAS } from "../config/providerModels.js";
import { createErrorResult, parseUpstreamError, formatProviderError } from "../utils/error.js";
import { HTTP_STATUS } from "../config/constants.js";
import { getExecutor } from "../executors/index.js";
import { refreshWithRetry } from "../services/tokenRefresh.js";
/**
* Build the embeddings request body for the target provider.
* Most OpenAI-compatible providers accept the same format.
* For providers that don't support embeddings natively (chat-only), we return an error.
*/
function buildEmbeddingsBody(model, input, encodingFormat) {
const body = {
model,
input
};
if (encodingFormat) {
body.encoding_format = encodingFormat;
}
return body;
}
/**
* Build the URL for the embeddings endpoint based on the provider.
*/
function buildEmbeddingsUrl(provider, credentials) {
switch (provider) {
case "openai":
return "https://api.openai.com/v1/embeddings";
case "openrouter":
return "https://openrouter.ai/api/v1/embeddings";
default:
// openai-compatible providers: use their baseUrl + /embeddings
if (provider?.startsWith?.("openai-compatible-")) {
const baseUrl = credentials?.providerSpecificData?.baseUrl || "https://api.openai.com/v1";
return `${baseUrl.replace(/\/$/, "")}/embeddings`;
}
// For other providers, attempt to use their base URL pattern with /embeddings path
return null;
}
}
/**
* Build headers for the embeddings request.
*/
function buildEmbeddingsHeaders(provider, credentials) {
const headers = { "Content-Type": "application/json" };
switch (provider) {
case "openai":
case "openrouter":
headers["Authorization"] = `Bearer ${credentials.apiKey || credentials.accessToken}`;
if (provider === "openrouter") {
headers["HTTP-Referer"] = "https://endpoint-proxy.local";
headers["X-Title"] = "Endpoint Proxy";
}
break;
default:
if (provider?.startsWith?.("openai-compatible-")) {
headers["Authorization"] = `Bearer ${credentials.apiKey || credentials.accessToken}`;
} else {
headers["Authorization"] = `Bearer ${credentials.apiKey || credentials.accessToken}`;
}
}
return headers;
}
/**
* Normalize the embeddings response to OpenAI format.
* Most OpenAI-compatible providers already return this format.
*/
function normalizeEmbeddingsResponse(responseBody, model) {
// Already in OpenAI format
if (responseBody.object === "list" && Array.isArray(responseBody.data)) {
return responseBody;
}
// Try to handle alternate formats gracefully
return responseBody;
}
/**
* Core embeddings handler — shared between Worker and SSE server.
*
* @param {object} options
* @param {object} options.body - Parsed request body { model, input, encoding_format }
* @param {object} options.modelInfo - { provider, model }
* @param {object} options.credentials - Provider credentials
* @param {object} [options.log] - Logger
* @param {function} [options.onCredentialsRefreshed] - Called when creds are refreshed
* @param {function} [options.onRequestSuccess] - Called on success (clear error state)
* @returns {Promise<{ success: boolean, response: Response, status?: number, error?: string }>}
*/
export async function handleEmbeddingsCore({
body,
modelInfo,
credentials,
log,
onCredentialsRefreshed,
onRequestSuccess
}) {
const { provider, model } = modelInfo;
// Validate input
const input = body.input;
if (!input) {
return createErrorResult(HTTP_STATUS.BAD_REQUEST, "Missing required field: input");
}
if (typeof input !== "string" && !Array.isArray(input)) {
return createErrorResult(HTTP_STATUS.BAD_REQUEST, "input must be a string or array of strings");
}
const encodingFormat = body.encoding_format || "float";
// Determine embeddings URL
const url = buildEmbeddingsUrl(provider, credentials);
if (!url) {
return createErrorResult(
HTTP_STATUS.BAD_REQUEST,
`Provider '${provider}' does not support embeddings. Use openai, openrouter, or an openai-compatible provider.`
);
}
const headers = buildEmbeddingsHeaders(provider, credentials);
const requestBody = buildEmbeddingsBody(model, input, encodingFormat);
log?.debug?.("EMBEDDINGS", `${provider.toUpperCase()} | ${model} | input_type=${Array.isArray(input) ? `array[${input.length}]` : "string"}`);
let providerResponse;
try {
providerResponse = await fetch(url, {
method: "POST",
headers,
body: JSON.stringify(requestBody)
});
} catch (error) {
const errMsg = formatProviderError(error, provider, model, HTTP_STATUS.BAD_GATEWAY);
log?.debug?.("EMBEDDINGS", `Fetch error: ${errMsg}`);
return createErrorResult(HTTP_STATUS.BAD_GATEWAY, errMsg);
}
// Handle 401/403 — try token refresh
if (
providerResponse.status === HTTP_STATUS.UNAUTHORIZED ||
providerResponse.status === HTTP_STATUS.FORBIDDEN
) {
const executor = getExecutor(provider);
const newCredentials = await refreshWithRetry(
() => executor.refreshCredentials(credentials, log),
3,
log
);
if (newCredentials?.accessToken || newCredentials?.apiKey) {
log?.info?.("TOKEN", `${provider.toUpperCase()} | refreshed for embeddings`);
Object.assign(credentials, newCredentials);
if (onCredentialsRefreshed && newCredentials) {
await onCredentialsRefreshed(newCredentials);
}
// Retry with refreshed credentials
try {
const retryHeaders = buildEmbeddingsHeaders(provider, credentials);
providerResponse = await fetch(url, {
method: "POST",
headers: retryHeaders,
body: JSON.stringify(requestBody)
});
} catch (retryError) {
log?.warn?.("TOKEN", `${provider.toUpperCase()} | retry after refresh failed`);
}
} else {
log?.warn?.("TOKEN", `${provider.toUpperCase()} | refresh failed`);
}
}
if (!providerResponse.ok) {
const { statusCode, message } = await parseUpstreamError(providerResponse, provider);
const errMsg = formatProviderError(new Error(message), provider, model, statusCode);
log?.debug?.("EMBEDDINGS", `Provider error: ${errMsg}`);
return createErrorResult(statusCode, errMsg);
}
let responseBody;
try {
responseBody = await providerResponse.json();
} catch (parseError) {
return createErrorResult(HTTP_STATUS.BAD_GATEWAY, `Invalid JSON response from ${provider}`);
}
if (onRequestSuccess) {
await onRequestSuccess();
}
const normalized = normalizeEmbeddingsResponse(responseBody, model);
log?.debug?.("EMBEDDINGS", `Success | usage=${JSON.stringify(normalized.usage || {})}`);
return {
success: true,
response: new Response(JSON.stringify(normalized), {
headers: {
"Content-Type": "application/json",
"Access-Control-Allow-Origin": "*"
}
})
};
}