mirror of
https://github.com/decolua/9router.git
synced 2026-05-08 12:01:28 +00:00
104 lines
3.7 KiB
JavaScript
104 lines
3.7 KiB
JavaScript
/**
|
|
* Responses API Handler for Workers
|
|
* Converts Chat Completions to Codex Responses API format
|
|
*/
|
|
|
|
import { handleChatCore } from "./chatCore.js";
|
|
import { convertResponsesApiFormat } from "../translator/helpers/responsesApiHelper.js";
|
|
import { createResponsesApiTransformStream } from "../transformer/responsesTransformer.js";
|
|
import { convertResponsesStreamToJson } from "../transformer/streamToJsonConverter.js";
|
|
|
|
/**
|
|
* Handle /v1/responses request
|
|
* @param {object} options
|
|
* @param {object} options.body - Request body (Responses API format)
|
|
* @param {object} options.modelInfo - { provider, model }
|
|
* @param {object} options.credentials - Provider credentials
|
|
* @param {object} options.log - Logger instance (optional)
|
|
* @param {function} options.onCredentialsRefreshed - Callback when credentials are refreshed
|
|
* @param {function} options.onRequestSuccess - Callback when request succeeds
|
|
* @param {function} options.onDisconnect - Callback when client disconnects
|
|
* @param {string} options.connectionId - Connection ID for usage tracking
|
|
* @returns {Promise<{success: boolean, response?: Response, status?: number, error?: string}>}
|
|
*/
|
|
export async function handleResponsesCore({ body, modelInfo, credentials, log, onCredentialsRefreshed, onRequestSuccess, onDisconnect, connectionId }) {
|
|
// Convert Responses API format to Chat Completions format
|
|
const convertedBody = convertResponsesApiFormat(body);
|
|
|
|
// Preserve client's stream preference (matches OpenClaw behavior)
|
|
// Default to false if omitted: Boolean(undefined) = false
|
|
const clientRequestedStreaming = convertedBody.stream === true;
|
|
if (convertedBody.stream === undefined) {
|
|
convertedBody.stream = false;
|
|
}
|
|
|
|
// Call chat core handler — force sourceFormat so streaming path knows this is a Responses API client
|
|
const result = await handleChatCore({
|
|
body: convertedBody,
|
|
modelInfo,
|
|
credentials,
|
|
log,
|
|
onCredentialsRefreshed,
|
|
onRequestSuccess,
|
|
onDisconnect,
|
|
connectionId,
|
|
sourceFormatOverride: "openai-responses"
|
|
});
|
|
|
|
if (!result.success || !result.response) {
|
|
return result;
|
|
}
|
|
|
|
const response = result.response;
|
|
const contentType = response.headers.get("Content-Type") || "";
|
|
|
|
// Case 1: Client wants non-streaming, but got SSE (provider forced it, e.g., Codex)
|
|
if (!clientRequestedStreaming && contentType.includes("text/event-stream")) {
|
|
try {
|
|
const jsonResponse = await convertResponsesStreamToJson(response.body);
|
|
|
|
return {
|
|
success: true,
|
|
response: new Response(JSON.stringify(jsonResponse), {
|
|
status: 200,
|
|
headers: {
|
|
"Content-Type": "application/json",
|
|
"Cache-Control": "no-cache",
|
|
"Access-Control-Allow-Origin": "*"
|
|
}
|
|
})
|
|
};
|
|
} catch (error) {
|
|
console.error("[Responses API] Stream-to-JSON conversion failed:", error);
|
|
return {
|
|
success: false,
|
|
status: 500,
|
|
error: "Failed to convert streaming response to JSON"
|
|
};
|
|
}
|
|
}
|
|
|
|
// Case 2: Client wants streaming, got SSE - transform it
|
|
if (clientRequestedStreaming && contentType.includes("text/event-stream")) {
|
|
const transformStream = createResponsesApiTransformStream(null);
|
|
const transformedBody = response.body.pipeThrough(transformStream);
|
|
|
|
return {
|
|
success: true,
|
|
response: new Response(transformedBody, {
|
|
status: 200,
|
|
headers: {
|
|
"Content-Type": "text/event-stream",
|
|
"Cache-Control": "no-cache",
|
|
"Connection": "keep-alive",
|
|
"Access-Control-Allow-Origin": "*"
|
|
}
|
|
})
|
|
};
|
|
}
|
|
|
|
// Case 3: Non-SSE response (error or non-streaming from provider) - return as-is
|
|
return result;
|
|
}
|
|
|