From 45731ae639b00eb0e207a031ca294dd265b7bd9f Mon Sep 17 00:00:00 2001 From: decolua Date: Wed, 22 Apr 2026 14:16:21 +0700 Subject: [PATCH] feat: add OpenCode Go provider and support for custom models - Introduced OpenCode Go provider with relevant configurations. - Enhanced model management by allowing users to add and delete custom models. - Updated UI components to support model selection for image types. - Adjusted sidebar visibility to include image media kinds. --- open-sse/config/providerModels.js | 45 +++ open-sse/config/providers.js | 5 + open-sse/executors/index.js | 3 + open-sse/executors/opencode-go.js | 51 +++ open-sse/handlers/imageGenerationCore.js | 320 ++++++++++++++++++ open-sse/services/model.js | 1 + public/providers/opencode-go.png | Bin 0 -> 16378 bytes .../media-providers/[kind]/[id]/page.js | 43 ++- .../providers/components/ModelsCard.js | 70 ++-- src/app/api/models/custom/route.js | 48 +++ src/app/api/v1/images/generations/route.js | 16 + src/lib/localDb.js | 28 ++ src/models/index.js | 3 + src/shared/components/Sidebar.js | 2 +- src/shared/constants/providers.js | 3 +- src/sse/handlers/imageGeneration.js | 152 +++++++++ tests/unit/image-generation.test.js | 320 ++++++++++++++++++ tests/vitest.config.js | 2 + 18 files changed, 1076 insertions(+), 36 deletions(-) create mode 100644 open-sse/executors/opencode-go.js create mode 100644 open-sse/handlers/imageGenerationCore.js create mode 100644 public/providers/opencode-go.png create mode 100644 src/app/api/models/custom/route.js create mode 100644 src/app/api/v1/images/generations/route.js create mode 100644 src/sse/handlers/imageGeneration.js create mode 100644 tests/unit/image-generation.test.js diff --git a/open-sse/config/providerModels.js b/open-sse/config/providerModels.js index b80c3b1c..b913360a 100644 --- a/open-sse/config/providerModels.js +++ b/open-sse/config/providerModels.js @@ -141,6 +141,18 @@ export const PROVIDER_MODELS = { { id: "deepseek/deepseek-chat", name: "DeepSeek Chat" }, { id: "deepseek/deepseek-reasoner", name: "DeepSeek Reasoner" }, ], + "opencode-go": [ // OpenCode Go subscription (API key) + { id: "kimi-k2.6", name: "Kimi K2.6" }, + { id: "kimi-k2.5", name: "Kimi K2.5" }, + { id: "glm-5.1", name: "GLM 5.1" }, + { id: "glm-5", name: "GLM 5" }, + { id: "qwen3.5-plus", name: "Qwen 3.5 Plus" }, + { id: "qwen3.6-plus", name: "Qwen 3.6 Plus" }, + { id: "mimo-v2-pro", name: "MiMo V2 Pro" }, + { id: "mimo-v2-omni", name: "MiMo V2 Omni" }, + { id: "minimax-m2.7", name: "MiniMax M2.7", targetFormat: "claude" }, + { id: "minimax-m2.5", name: "MiniMax M2.5", targetFormat: "claude" }, + ], oc: [ // OpenCode // { id: "nemotron-3-super-free", name: "Nemotron 3 Super" }, // { id: "qwen3.6-plus-free", name: "Qwen 3.6 Plus" }, @@ -192,6 +204,10 @@ export const PROVIDER_MODELS = { { id: "tts-1", name: "TTS-1", type: "tts" }, { id: "tts-1-hd", name: "TTS-1 HD", type: "tts" }, { id: "gpt-4o-mini-tts", name: "GPT-4o Mini TTS", type: "tts" }, + // Image models + { id: "gpt-image-1", name: "GPT Image 1", type: "image" }, + { id: "dall-e-3", name: "DALL-E 3", type: "image" }, + { id: "dall-e-2", name: "DALL-E 2", type: "image" }, ], anthropic: [ { id: "claude-sonnet-4-20250514", name: "Claude Sonnet 4" }, @@ -219,6 +235,10 @@ export const PROVIDER_MODELS = { { id: "gemini-embedding-001", name: "Gemini Embedding 001", type: "embedding" }, { id: "text-embedding-005", name: "Text Embedding 005", type: "embedding" }, { id: "text-embedding-004", name: "Text Embedding 004 (Legacy)", type: "embedding" }, + // Image models (Nano Banana) + { id: "gemini-3.1-flash-image-preview", name: "Gemini 3.1 Flash Image (Nano Banana 2)", type: "image" }, + { id: "gemini-3-pro-image-preview", name: "Gemini 3 Pro Image (Nano Banana Pro)", type: "image" }, + { id: "gemini-2.5-flash-image", name: "Gemini 2.5 Flash Image (Nano Banana)", type: "image" }, ], openrouter: [ // Embedding models @@ -233,6 +253,11 @@ export const PROVIDER_MODELS = { { id: "openai/gpt-4o-mini-tts", name: "GPT-4o Mini TTS", type: "tts" }, { id: "openai/tts-1-hd", name: "TTS-1 HD", type: "tts" }, { id: "openai/tts-1", name: "TTS-1", type: "tts" }, + // Image models + { id: "openai/dall-e-3", name: "DALL-E 3 (via OpenRouter)", type: "image" }, + { id: "openai/gpt-image-1", name: "GPT Image 1 (via OpenRouter)", type: "image" }, + { id: "google/imagen-3.0-generate-002", name: "Imagen 3 (via OpenRouter)", type: "image" }, + { id: "black-forest-labs/FLUX.1-schnell", name: "FLUX.1 Schnell (via OpenRouter)", type: "image" }, ], glm: [ { id: "glm-5.1", name: "GLM 5.1" }, @@ -256,6 +281,8 @@ export const PROVIDER_MODELS = { { id: "MiniMax-M2.7", name: "MiniMax M2.7" }, { id: "MiniMax-M2.5", name: "MiniMax M2.5" }, { id: "MiniMax-M2.1", name: "MiniMax M2.1" }, + // Image models + { id: "minimax-image-01", name: "MiniMax Image 01", type: "image" }, ], blackbox: [ { id: "gpt-4o", name: "GPT-4o" }, @@ -424,6 +451,24 @@ export const PROVIDER_MODELS = { // TTS entries are loaded from ttsModels.js via buildTtsProviderModels() ...buildTtsProviderModels(), + + // Image providers + nanobanana: [ + { id: "nanobanana-flash", name: "NanoBanana Flash", type: "image" }, + { id: "nanobanana-pro", name: "NanoBanana Pro", type: "image" }, + ], + sdwebui: [ + { id: "stable-diffusion-v1-5", name: "Stable Diffusion v1.5", type: "image" }, + { id: "sdxl-base-1.0", name: "SDXL Base 1.0", type: "image" }, + ], + comfyui: [ + { id: "flux-dev", name: "FLUX Dev", type: "image" }, + { id: "sdxl", name: "SDXL", type: "image" }, + ], + huggingface: [ + { id: "black-forest-labs/FLUX.1-schnell", name: "FLUX.1 Schnell", type: "image" }, + { id: "stabilityai/stable-diffusion-xl-base-1.0", name: "SDXL Base 1.0", type: "image" }, + ], }; // Helper functions diff --git a/open-sse/config/providers.js b/open-sse/config/providers.js index c7c9ee82..3c5dabdc 100644 --- a/open-sse/config/providers.js +++ b/open-sse/config/providers.js @@ -337,6 +337,11 @@ export const PROVIDERS = { headers: { "x-opencode-client": "desktop" }, noAuth: true }, + "opencode-go": { + baseUrl: "https://opencode.ai/zen/go/v1/chat/completions", + format: "openai", + headers: {} + }, "grok-web": { baseUrl: "https://grok.com/rest/app-chat/conversations/new", format: "grok-web", diff --git a/open-sse/executors/index.js b/open-sse/executors/index.js index f1e175f4..b7ca8a15 100644 --- a/open-sse/executors/index.js +++ b/open-sse/executors/index.js @@ -9,6 +9,7 @@ import { CursorExecutor } from "./cursor.js"; import { VertexExecutor } from "./vertex.js"; import { QwenExecutor } from "./qwen.js"; import { OpenCodeExecutor } from "./opencode.js"; +import { OpenCodeGoExecutor } from "./opencode-go.js"; import { GrokWebExecutor } from "./grok-web.js"; import { PerplexityWebExecutor } from "./perplexity-web.js"; import { DefaultExecutor } from "./default.js"; @@ -27,6 +28,7 @@ const executors = { "vertex-partner": new VertexExecutor("vertex-partner"), qwen: new QwenExecutor(), opencode: new OpenCodeExecutor(), + "opencode-go": new OpenCodeGoExecutor(), "grok-web": new GrokWebExecutor(), "perplexity-web": new PerplexityWebExecutor(), }; @@ -56,5 +58,6 @@ export { VertexExecutor } from "./vertex.js"; export { DefaultExecutor } from "./default.js"; export { QwenExecutor } from "./qwen.js"; export { OpenCodeExecutor } from "./opencode.js"; +export { OpenCodeGoExecutor } from "./opencode-go.js"; export { GrokWebExecutor } from "./grok-web.js"; export { PerplexityWebExecutor } from "./perplexity-web.js"; diff --git a/open-sse/executors/opencode-go.js b/open-sse/executors/opencode-go.js new file mode 100644 index 00000000..c75a4724 --- /dev/null +++ b/open-sse/executors/opencode-go.js @@ -0,0 +1,51 @@ +import { BaseExecutor } from "./base.js"; +import { PROVIDERS } from "../config/providers.js"; + +// Models that use /zen/go/v1/messages (Anthropic/Claude format + x-api-key auth) +const CLAUDE_FORMAT_MODELS = new Set(["minimax-m2.5", "minimax-m2.7"]); + +const BASE = "https://opencode.ai/zen/go/v1"; + +// Kimi (Moonshot) requires reasoning_content on assistant tool_call messages when thinking is on. +// OpenAI-format clients don't send it -> upstream 400. Inject a non-empty placeholder. +const KIMI_REASONING_PLACEHOLDER = " "; + +export class OpenCodeGoExecutor extends BaseExecutor { + constructor() { + super("opencode-go", PROVIDERS["opencode-go"]); + } + + // buildUrl runs before buildHeaders in BaseExecutor.execute, cache model here + buildUrl(model) { + this._lastModel = model; + return CLAUDE_FORMAT_MODELS.has(model) + ? `${BASE}/messages` + : `${BASE}/chat/completions`; + } + + buildHeaders(credentials, stream = true) { + const key = credentials?.apiKey || credentials?.accessToken; + const headers = { "Content-Type": "application/json" }; + + if (CLAUDE_FORMAT_MODELS.has(this._lastModel)) { + headers["x-api-key"] = key; + headers["anthropic-version"] = "2023-06-01"; + } else { + headers["Authorization"] = `Bearer ${key}`; + } + + if (stream) headers["Accept"] = "text/event-stream"; + return headers; + } + + transformRequest(model, body) { + if (!model?.startsWith?.("kimi-") || !body?.messages) return body; + const messages = body.messages.map(m => { + if (m?.role === "assistant" && Array.isArray(m.tool_calls) && !("reasoning_content" in m)) { + return { ...m, reasoning_content: KIMI_REASONING_PLACEHOLDER }; + } + return m; + }); + return { ...body, messages }; + } +} diff --git a/open-sse/handlers/imageGenerationCore.js b/open-sse/handlers/imageGenerationCore.js new file mode 100644 index 00000000..2189f5f7 --- /dev/null +++ b/open-sse/handlers/imageGenerationCore.js @@ -0,0 +1,320 @@ +import { createErrorResult, parseUpstreamError, formatProviderError } from "../utils/error.js"; +import { HTTP_STATUS } from "../config/runtimeConfig.js"; +import { refreshWithRetry } from "../services/tokenRefresh.js"; +import { getExecutor } from "../executors/index.js"; + +// Image provider configurations +const IMAGE_PROVIDERS = { + openai: { + baseUrl: "https://api.openai.com/v1/images/generations", + format: "openai", + }, + gemini: { + baseUrl: "https://generativelanguage.googleapis.com/v1beta/models", + format: "gemini", + }, + minimax: { + baseUrl: "https://api.minimaxi.com/v1/images/generations", + format: "openai", + }, + openrouter: { + baseUrl: "https://openrouter.ai/api/v1/images/generations", + format: "openai", + }, + nanobanana: { + baseUrl: "https://api.nanobananaapi.ai/api/v1/nanobanana/generate", + format: "nanobanana", + }, + sdwebui: { + baseUrl: "http://localhost:7860/sdapi/v1/txt2img", + format: "sdwebui", + }, + comfyui: { + baseUrl: "http://localhost:8188", + format: "comfyui", + }, + huggingface: { + baseUrl: "https://api-inference.huggingface.co/models", + format: "huggingface", + }, +}; + +/** + * Build image generation URL + */ +function buildImageUrl(provider, model, credentials) { + const config = IMAGE_PROVIDERS[provider]; + if (!config) return null; + + switch (provider) { + case "gemini": { + const apiKey = credentials?.apiKey || credentials?.accessToken; + const modelId = model.replace(/^models\//, ""); + return `${config.baseUrl}/${modelId}:generateContent?key=${encodeURIComponent(apiKey)}`; + } + case "huggingface": + return `${config.baseUrl}/${model}`; + default: + return config.baseUrl; + } +} + +/** + * Build request headers + */ +function buildImageHeaders(provider, credentials) { + const headers = { "Content-Type": "application/json" }; + + if (provider === "gemini") { + return headers; + } + + if (provider === "openrouter") { + headers["Authorization"] = `Bearer ${credentials?.apiKey || credentials?.accessToken}`; + headers["HTTP-Referer"] = "https://endpoint-proxy.local"; + headers["X-Title"] = "Endpoint Proxy"; + return headers; + } + + if (provider === "huggingface") { + headers["Authorization"] = `Bearer ${credentials?.apiKey || credentials?.accessToken}`; + return headers; + } + + if (credentials?.apiKey || credentials?.accessToken) { + headers["Authorization"] = `Bearer ${credentials.apiKey || credentials.accessToken}`; + } + + return headers; +} + +/** + * Build request body based on provider format + */ +function buildImageBody(provider, model, body) { + const { prompt, n = 1, size = "1024x1024", quality, style, response_format } = body; + + switch (provider) { + case "gemini": + return { + contents: [{ parts: [{ text: prompt }] }], + generationConfig: { + responseModalities: ["TEXT", "IMAGE"], + }, + }; + + case "sdwebui": { + const [width, height] = size.split("x").map(Number); + return { + prompt, + width: width || 512, + height: height || 512, + steps: 20, + batch_size: n, + }; + } + + case "nanobanana": { + const sizeMap = { + "1024x1024": "1:1", + "1024x1792": "9:16", + "1792x1024": "16:9", + }; + return { + prompt, + type: "TEXTTOIAMGE", + numImages: n, + image_size: sizeMap[size] || "1:1", + }; + } + + default: + // OpenAI-compatible format + const requestBody = { model, prompt, n, size }; + if (quality) requestBody.quality = quality; + if (style) requestBody.style = style; + if (response_format) requestBody.response_format = response_format; + return requestBody; + } +} + +/** + * Normalize response to OpenAI format + */ +function normalizeImageResponse(responseBody, provider, prompt) { + // Already in OpenAI format + if (responseBody.created && Array.isArray(responseBody.data)) { + return responseBody; + } + + const timestamp = Math.floor(Date.now() / 1000); + + switch (provider) { + case "gemini": { + const parts = responseBody.candidates?.[0]?.content?.parts || []; + const images = parts + .filter((p) => p.inlineData?.data) + .map((p) => ({ b64_json: p.inlineData.data })); + return { + created: timestamp, + data: images.length > 0 ? images : [{ b64_json: "", revised_prompt: prompt }], + }; + } + + case "sdwebui": { + const images = Array.isArray(responseBody.images) + ? responseBody.images.map((img) => ({ b64_json: img })) + : []; + return { created: timestamp, data: images }; + } + + case "nanobanana": { + if (responseBody.image) { + return { + created: timestamp, + data: [{ b64_json: responseBody.image, revised_prompt: prompt }], + }; + } + return { created: timestamp, data: [] }; + } + + case "huggingface": { + // HuggingFace returns binary image data + return responseBody; + } + + default: + return responseBody; + } +} + +/** + * Core image generation handler + * @param {object} options + * @param {object} options.body - Request body { model, prompt, n, size, ... } + * @param {object} options.modelInfo - { provider, model } + * @param {object} options.credentials - Provider credentials + * @param {object} [options.log] - Logger + * @param {function} [options.onCredentialsRefreshed] - Called when creds are refreshed + * @param {function} [options.onRequestSuccess] - Called on success + * @returns {Promise<{ success: boolean, response: Response, status?: number, error?: string }>} + */ +export async function handleImageGenerationCore({ + body, + modelInfo, + credentials, + log, + onCredentialsRefreshed, + onRequestSuccess, +}) { + const { provider, model } = modelInfo; + + if (!body.prompt) { + return createErrorResult(HTTP_STATUS.BAD_REQUEST, "Missing required field: prompt"); + } + + const url = buildImageUrl(provider, model, credentials); + if (!url) { + return createErrorResult( + HTTP_STATUS.BAD_REQUEST, + `Provider '${provider}' does not support image generation` + ); + } + + const headers = buildImageHeaders(provider, credentials); + const requestBody = buildImageBody(provider, model, body); + + log?.debug?.("IMAGE", `${provider.toUpperCase()} | ${model} | prompt="${body.prompt.slice(0, 50)}..."`); + + let providerResponse; + try { + providerResponse = await fetch(url, { + method: "POST", + headers, + body: JSON.stringify(requestBody), + }); + } catch (error) { + const errMsg = formatProviderError(error, provider, model, HTTP_STATUS.BAD_GATEWAY); + log?.debug?.("IMAGE", `Fetch error: ${errMsg}`); + return createErrorResult(HTTP_STATUS.BAD_GATEWAY, errMsg); + } + + // Handle 401/403 — try token refresh + const executor = getExecutor(provider); + if ( + !executor?.noAuth && + (providerResponse.status === HTTP_STATUS.UNAUTHORIZED || + providerResponse.status === HTTP_STATUS.FORBIDDEN) + ) { + const newCredentials = await refreshWithRetry( + () => executor.refreshCredentials(credentials, log), + 3, + log + ); + + if (newCredentials?.accessToken || newCredentials?.apiKey) { + log?.info?.("TOKEN", `${provider.toUpperCase()} | refreshed for image generation`); + Object.assign(credentials, newCredentials); + if (onCredentialsRefreshed && newCredentials) { + await onCredentialsRefreshed(newCredentials); + } + + try { + const retryHeaders = buildImageHeaders(provider, credentials); + const retryUrl = provider === "gemini" ? buildImageUrl(provider, model, credentials) : url; + + providerResponse = await fetch(retryUrl, { + method: "POST", + headers: retryHeaders, + body: JSON.stringify(requestBody), + }); + } catch (retryError) { + log?.warn?.("TOKEN", `${provider.toUpperCase()} | retry after refresh failed`); + } + } else { + log?.warn?.("TOKEN", `${provider.toUpperCase()} | refresh failed`); + } + } + + if (!providerResponse.ok) { + const { statusCode, message } = await parseUpstreamError(providerResponse); + const errMsg = formatProviderError(new Error(message), provider, model, statusCode); + log?.debug?.("IMAGE", `Provider error: ${errMsg}`); + return createErrorResult(statusCode, errMsg); + } + + let responseBody; + try { + // HuggingFace returns binary image data + if (provider === "huggingface") { + const buffer = await providerResponse.arrayBuffer(); + const base64 = Buffer.from(buffer).toString("base64"); + responseBody = { + created: Math.floor(Date.now() / 1000), + data: [{ b64_json: base64 }], + }; + } else { + responseBody = await providerResponse.json(); + } + } catch (parseError) { + return createErrorResult(HTTP_STATUS.BAD_GATEWAY, `Invalid response from ${provider}`); + } + + if (onRequestSuccess) { + await onRequestSuccess(); + } + + const normalized = normalizeImageResponse(responseBody, provider, body.prompt); + + log?.debug?.("IMAGE", `Success | images=${normalized.data?.length || 0}`); + + return { + success: true, + response: new Response(JSON.stringify(normalized), { + headers: { + "Content-Type": "application/json", + "Access-Control-Allow-Origin": "*", + }, + }), + }; +} diff --git a/open-sse/services/model.js b/open-sse/services/model.js index 1b3d96ba..ebe8d90b 100644 --- a/open-sse/services/model.js +++ b/open-sse/services/model.js @@ -13,6 +13,7 @@ const ALIAS_TO_PROVIDER_ID = { kmc: "kimi-coding", cl: "cline", oc: "opencode", + ocg: "opencode-go", // TTS providers el: "elevenlabs", // API Key providers diff --git a/public/providers/opencode-go.png b/public/providers/opencode-go.png new file mode 100644 index 0000000000000000000000000000000000000000..2e709e1c49db5064be4dc445a3d6e3fb4407c1da GIT binary patch literal 16378 zcmYMbbyyp|7d~2i_y!wZ913H&6}K^9jIuF-;W~y(af;g*4jb+|hQpBJE`#B&#f!IS zfdcpQz4!P0Zl34lJxSgqe?*?-Bq#DMM2(z;fdl{mkZY*F)cFqs|0j$2Y~ng=HmhY5jFt)|HFa&hyM%l z|H=RA{C|j-i}Qcj|K;W4{$F-rF5dsI|F2p)UhC|CKC0E z)7nk?;9y={&6m3DrtzQzvC=gW4|hT5m8jsD2)4(=IpeS?*kjB>3GOavosEkKznG;?6*oE1|1sk|jn z8LQ{Tx5ph1_bfB!56=jE_DU4)_V|o0JmHHnZ0;Yi$5*{uI;xtPrt}2Zrd#K#FFvBENxtf{96e@ zIwe6DDgaOojKqtT_k)eL9X@X1*{6o~{5B1)q0y&J^jG)XJ$ZsWFuwk)dynOalRX_e za!xW_ShKXra=Z>uhhr_s9|kSr|tH|1E#rWf^BQQ7?Z+n}6Vc#iK$kQ;fw_udQF zZB4I|>@$--ev{taV3{<5d_U=?Ds}{q)mr5)vC~p^=C$L7qe*RIT`IE&> zZUu~FiaSP}rYnOc8|hHK$8`D{a#uM~y%%l7w395(v@;;Uyu`UgW1OlVwyLXo5Uutp zioE=w*^7jAaR&`9>6iL)2;-b1^hyw_^1UcCxvO&W0W~$?_#-3yk@s=1Eslq`kq>CU zZI$OhwBc9oS8vnZD56~??J*yA<(pPGGN?_y`i=;Jmxu{=C&!AB;I2)+k4Ecc*?9{k z-a%Rd&whKc4>Kd#KAln0UUw4cRs2h-oUV)$t~JC9GNZEnmVFS`UGAahDrCyNE|ag-P^s+|KwvM1~6Km?gmt z;}aI&1@ZW13*YA=B_rI~*c31#jsoGMM!-DTSACc*ykmx9zd++m+G?PEk~H=_d+w7d z?P^ViBs68HK|`dqE2COc*&bLG5i2JC6>B8d@3dF2!2(++HNlip!Vq{7S3zly5p(UA z^=mwcmW$61MKE|Bp2dkNl3yVSC|lu2@o~jxXcT9(s4;VDAsOATR^k<_x?H3=8nGev z@S^`x>(I2<-9fm9i%>*Zo1bL)y{HOgzEt}ZLdj3TzyjXL!N^@2c6zqbAIN0#RUD}m z1>E_r-L9J8sk9hnjWwQQZ-;%3ebEz-pKd^_kEji9IR6NJ>w$J#;l0k%&yp1+=oDf_ z9a*fLQi43vUCt%{yfZ(^m!de$9=@cU?aa82kQ8nVI7pPdGlTN`M;hZ5>73xNFZng$ zlrB6qayA^d6->UbdYF&9n=ik8da%@fwA8L4NTwi&Aocb}#+jf!g(KSP@j2f5Oqbxx z!W&9>dNa$!bj*3vzbI^m1WWAFI5~BOr|>7#;DBRtWIVjh~=nO&)cusP80wm z5XAn&6Rt~TdQg1-PaDBgp~mr@r%QS?{61Ywt?qx#z6*Kc{|X*>_s`-cjL3t|JWfRZ zQ~CRdx*VZVF1?FC*XE>>{hwwxQhB9)>YYD0#~7TX(+j=-;|u%J{*SgzX>y|ceU3NP zFrmF4KsiSiRl_ew@hQ1VN{49nlN~y+UtvCV5wLcDw06(HyW0J;ijat(ku6E{?NT+g zP@^I1us~NEZ{wxNR~E)a{~cGVZY==pXRDE~vKEJH{S+g~t@@+Y<{Wm0y)5j$MPyiHw8GX`p&rM*#}f518_PX2!dW4QLw$y_B$f~D zC)m*w>~k2D2;<0m$6<8g3}zQ$y?_dRl6gwr@-wEV&As@0xjc+Yv%w;#)8*IptgY<# zpGlI+)hz>* zo)&A5jHIErR+dpyOev2~hjFS->}x@0O14_fgn|>VFw_XgE7God-M|LvURNXfc<{5Km|!45pO!ZB;JM-+?E~A` zyd}udyqPxa&egV^tc^z$OoGt{B5VZSf$5ht%RO^Fw6TZ^$ILSeeLqeU@)qY5Fs7sd z%RF#oaD%*h&KscA_@H}vrhpGzG75fU+LZJ)uwvkr=svy@LG^}RG@$+Q{N$d5S zFUm@hDanQn0u3>g5a}}ktkmdgeDJkxt>d+#pT)+1m#eSO+~^EuEiidDodg7d9W)N<+@We+HEMmL(q z=i;)l1gx zY;#`DWYX|%zv_N}re*a)V9abgfq^_FGR}LnmD=taG`+=>!IHNKFe%MBzHh%R#1!bm z0yw@|&~lIx!avl-Q-mv%PeRS9c_K+X^-PaN)*kNG0*Z2UL2yB2cKR9aZLDW&$Yy1G z4n z2Wad<45X8uRWw>*(zQYh7$`oKIBj!I`-JGn7T^u6!jOlWrm=zzngqX|5p~Rjd=l3j z9w%idbC~m%XNb`9h&q>IRT%=iZU!iL2F{s&1>hoFzJfBOCTJpEirO-Ji2iso$YUw; zDBVb*E?C^3zX-(hSax$94VX|etC%#CDA%o?p2ifObUpf{jt-;8N*?GPH9j%zNISj4|~$n*VJVE1O_ zib374sc)%a)^wI8zKAOZb@FB52t1OR|c%td!%TO{roa&SVpuu*?scFE=iQBMCF2PCfItO z$yqR1P%L1fHlr~Z>%Tl|kV%?Y~1cgVhFlZ4&xKkVAn z{~poAW%?0yDkyq!jIryM4?C=})|QgH@K5Z>2sWWHTdYDcfpvI|h8nuAiQ8VH$6HWp zM{cfi4{qySC9~i)5sx@hJ^5+ax^Pk0T3PKik+YYKq@+7ii`7LyY+u9ijCtlMv2Zk% z$ATkrn3#pbwz}!}CfAB9^7evL%}vd|xbNDdtA)IqN5B z3qHlh&efmP%)JtcqL+@amaB!BM!GoGoTvqQIkxVHWyQGcNMQhfa3eG|waOFxXqdel z<)jBmui%)+94^#j*eFR_#px#ZbaalEUS=}1w#A0Mm zBlh*SzQ+#hs4=rV;Ud6{+A5RTcu87{$+1@=g&LXso^H`kJ~p&?-4m&gd9Zp{6)ata$Q_ zhOJ??APMXlslvX_N-wlxjk~U)KppIBL4OjiB4$9b!);{0%UtenR5blmqE@Vii&hc& zmD)reu3WYK`K-O4e$D(jgq*Y!eJ>A%fev7V zWw}LlFVFZE4eWHl9)iyxD<`*$QZ|(n##Z57h2zK`;Vg+!v2*=91_OEOTr8ztxtWxwe_*a5x{TM8x8z>lW;I8jb zI!XC20^ZoH(?RHYeDdKY?OmH4+bFmb>~N1VmD#iaVnZ#-M8~-%Khu7vSNK`hZRPPx zU$$8CkJHFgMxC0*4x`4=6Ts1-ePfLkQ5b8(FLnHCGB zP&3``s%)Tk%uv{sWqMwS*vaSPURjniw;=0$bC;Xbfw@RZsFvFS!c+iWfD_{g2llT2 zTOJML~_YIy$7?1*|8w z9wdiJA-mVSTJR82k3OcGst$$6i+%QV$%hKA5GkO7%G54W{yu#aaK8H`_Nc?M`st6> z=4y|trpWFzVrH1_+l-eYXI?{^#;zC7OwiACwM9YZ6%wtI$cx4HvqlatTY>M{uZ(1) zYbHLeM#%I{GR~E(cYkDJt|R_}mq~)l-L>G50K)l(o_MlyDcl~?LaOG5zx&IumrF&9 z|J!?Q^bbjl$%LSG>%-ki$E1RKJ*kUplYDPVro!6MaQoqK&|$)acX5%JCMq4nimb{MlYwxC;1Q#p~rzZ<^*8 zsL`_YTPrLi_I8Cg2aqlhh$`UQzJxSnoFz;m+e^KI4YC4cZ_@ zVwspyk5bSJlKxJ10oa8s(11# zC+1JI6XQI%#E@VfA)NV{rBEVGtIobO_l-K3L1>rZ&R~H(%)Qxmhs8I8QxSTMCAL=>< zhQa3~A|#FuC={<^z%7Jqc~QLfSE3 z64)ft&e@(N;F%AT=7ViUYn;}Ydn$T-c-6zEu;L%gOXB3@qZ4KZU3OUS*2_#8sd8O- z>E>-7i>LWo>chB-Q_(3l*;c{B%c`%^dV#C2Ddyb{O>11NZl2ck^x+f~|XNC%dm4&G3Y8H?o| z^7+iY8~?uU)i9fx&fy-Hv-_udG_J-37oOOVpy!qC_G!fwD%|C2Nfl@l4INOQL>m1# z;`of8dT7BL_JbnUfoOmb0nDvRwyWtH|1oDar}eD^W2rLVD5(A`YLlGN}7z=1$)IazLcNgJ_L0#6ZE*)1VQlkj-w@x>e*y$>Gyt5Wta z05oa;H0)T08NdS-Dukp&e)gczQVX!fSQD^of6aV2>_oQ_2O!tV7@Duk{3)9 z((V!rb}Icnh}NK#!0^Wgp7*P2o8{Z$ke9aXR>CmBhJnLk#P@&lY*H_CN;l+}dUDv(4 z8tBw0Z|b^%TeE+u>Ob~Kpye(^e64#DKa+{y?f{J&;XrsO^!VA~j}_LemxlZ6FRkZK z)Zt9~<}C!qth6Z`Rm!3mZYwABq=oWpGJIj+PAA3l^WKkq3fGCEHy*#>UlCVnlRl0k z3F`j%#+WQxp`G$3Zf=lyF3WLpjmxw%%@WBETTW_4-scKVrWi*bqWM@4$YU<<8pnh$ zA#GHTq`dx>wPWV#-0-))QOezjSufFje_5bDYrTArxfBi~al$x*j2lHAX+b%h(Keuo znJ9K#e%XBvP=x+lSE3)ODO;51R@-SDc|thcb>aY#?V>7i7b%KXF_ z5D=(Ob~!WOiefF3EFRz>wPb}Hg^;&i=O37vS&n7+FJ)ya&{lP)M!XP!3JXkI{gw%= zj*LrBzK&Pe^MdAI<+?6M#pe`k4rc~Bw7Ao?q`f*N06}8xv%PIXGw~n{7Q>iw8UQb%}an~2U*MwyD$lbO8|4xVfIWX%IVzlcr|Ccnsj+ED<6@v>0VU%a$9)C#v zrVy6|!}~@$fVzY!KWzb6e$b_i-k8S8@in@uv^o{b*LEJagp%tYK^p9u=QmT(?Rcqn ziPVy(5zYUTmcv%jpJ zZ_7(*!Om4Z9US}odZge3afXx~l;cTcE+i9}#UnmpP-b`NQ{P5=R?OR292tNAcG9>p zSqh!b{=lqMXBN^Orl#lak0df2%s&$k6-lx5#c6E3PhS3CqBTducaBN+$Q)CjJ+-DQzPP zl)oqeS(2rfs&&Pat3QnUhQ78VdL(rht?m;8wKdI7mc8{&@B$Ve5PfiC!a3YOdiWoM z|0t65jl6yn)(KFVA*FKEV7aojVjNe`f#tG!0~U?t%}tH(6x zJ2PmAIoh^4Z-yvjzeD`79C)$w-4nykGfFNr!L3O7DuRu(`e;yub5%8-B8hDE(I8-~ z@l{J>1scD#d5XZLl|<1{dq>_c&iTP_?CgqoekxwJ{nrQk?J}2@VF5x-9#y z#FVg=5SiEAIwQSq{{%CzWC{Ir%CnwA$T! zy1p$FBdz!(=~a8SE;+1dDL3i+2Y8Up3n+s9oA#WoA$$3~p&e}d6ZEKEr-H!4ynFo1 ztJaQA-j#Z_Knt@+Ho??=FJ=lnL`-eweco}uh0e$IIc1`m%e%#AB{tn^V{v1WRgIGd|cv&bj=Ts%(V70v(O?mrAZb!5fiTDyYk9)LE-V}VzwC2vbMT!)v&D}uI zuB!P>L$M~XF4p&f3D*vbBsNCpoMwKvqQ#v8wUymjL?ZLlF$3Ar@boyqDvuGPl#3ma z8vJ`ntD+u>Mjm%E6gl7DU38uL$Yj?ot?Gm0Dr=%_hfd^vt>Ul623q~+CYLoppEu|b zJ9L7dR>2w0fD<*t$$NS#`)>?*mipL7Is$JPKTE&kcby%C{ZKL%;vVsSEag|2y9;Hn@pU5esKo93bG8Q|BaoxSg(ys5^ zj(?Ixek|i8ui)!&(c_xsq$==hO?EiD5gQTy_R}+P`1vPx%s0}2tNDUzDAC_9HIVr< zbtVn`?Bkfms^DAIQcyI6Gcc%GXqL^%cwr(#<)-!B!I#_sN>$I$)=ni*);-JXv+3Nb zbN{D22cHb~Js)i=OLG?N5TEaOrSB?ulJ;pk-cOdC(X;&H_a+>e_mnpS{(!T34lj$f zPX9iP#tC0#Ms8UEf$@E2vX@Ai5x29sE&%u|_*aIh3pZM1Xny`2sVzaCm11(=IrLEJ z?`H5C_etHx#|`B4^C(R+nIEOjz%rddW#?^VRFeNtPk(!vhI3I(lk@X2f1DS5O(Z0< z(2BgR-yRDY8hV2M7RB7SMKmHm4KNGQPixVyQE2I)m9Cer@ z-$9)3v5M9m#kCB_cXEM0NfAaC+tN?In)QQ-dtb2tbCC(eZ!C&&T?AgpzoF|prKFJe zzYMVX47+Dgi=r%`StJi2)^{wudEKknO5!xX9qKt(#yb*0k&~J!=u*GVwB8->U9={i zsz2WW`(Qf82!n+lsAsinz88;B$*gH9t`J@C9shbDqtfy`wxh3iw5H6P+wG1@a4R~+ z{teN*#v>hVGTx->8jP!unUDSrzeQt#5k5Uy^I;@ zQS)oy_=VM3^<5dNPwf}{&SHSya4Xppu%(e31PKJz1qTV+blEi~|2?aXvp>!ln4SkhGz3XCU$sXd z9>r@{v^E!x#Uh*h$-J2Ra$s_x?U9Yi@u~5FPc&tI=8p5-?@;9h6Oi9WgHX@p(OP&^ zEv^q$1>!Rg3ow|x>aJ1_7qCtkpA9&-<=#_Vcxr{*k^?kxuhj;OGRygKXXT`>%XmEF zG?SsH!V$_KlJTky7Gh>ipQQ~p41y4#wAb+DmTiwXh$fp))buD59|FO-YB!;#r zI$!F4PT>3H&Q_#*R*D7-&pJzx(P=-9hKU3mbXlQYMjXapWuewn3Dqb&<*PdIoFUIE z;MpbQxh$i|lwYlEdLMMiwM%BDjLs3~)Sihbj_@#RX>m3Qvku+VMivcJ?fSWcnJX9V zZUGjyog=9j3j&s6V6^u(2a;=O!@Q}U@~SbF+DlSX@C_xg#~887SX z3g|1*pu;XhjWa}@TG$0V-dOdKEa(f~mPF(eAOT4>^9wmJZ56kz9Q0H)wHv)r5*L^) zbNe<*-)?;G>mj~_ruh7e$>C>ijcZRH-f`@d3zEECvLGp%VSdiL45 zQcpA0SZ8<9G}N_IOz9O-H)0?6f&X>ZSN& zfX~WzYd#isc7tow)y&`ubHih9cv`{(lWS$j3FMjuU>eQqs_|faSU28G_JqXM-7oK}y%({# z^m-c&@&>R=weE>|8pstxEH2EHK>ai|#8i{opU`7}aR5RIqR!tv7NxLyf#fodS&Gxc z3dqY5WtdSs1V$PB(KPyL@eZU&{G!huD#*;1nlrj{^e={j)a6ZSQnoV&^w4t091laT zIFFk(PBM*-w0VUkFtrnUccp6r(WJCS&O|!o>7HLk=;~wZDB70H@CMCr3FG~_%-?`4 zphP5YZiEdDXb2&U;$uEae`LM9H-hp-1BTUlnvJlpR+RQ3)>Aw~z_-(*e5L1QC9j|TS=pklMHhDM7W;Cy;SVX-{~bP#G5 z?U1F^-gCjwL*TbJ@3HvL8eElQVuU@!`>6vn8A(Tu#3U{YTPCb6x}##P?b`6p!%f*R zdi|7ZZnU}fo_7mv8!}2o{~R*U*PA!@Foj!kTxp0F%n!?ybiwz6-;H4#5dOolFXb`w z@S&{CNR1zIC5X#ql8u2~9MYZXS5HQJHCV7(m-B;NIk=g*!EQQo|Mc~O?26JtV^vc* z_Z!$u?Yl1YjM}I>!(zZpVr$*}M&F-he`4_xcV?d7->0RGqqF&C%MVj0K4DE!4@q_e~~tfi|dsF?4W!CbM= zqxB&+(>r`^pIs$^XQO=;Bla)a6mX`$gf_{?+KkKu3^3S~T|;Hv4`h`8K_rqYO-|q1 z{#~@y9K|uT$r|PC#_o5V^vICyV@1A8infbM2@k*+OzV=H_miaI6dQZRR2Xyg8iQCr zn~oYu`KNmJp5(orJb61h;DOD8xT6e-0`9bo5D5TRhR33wQ;3sCI}{_OOjNB+%etH~ z<-(xZKw*x3?Ka@dwgCRg!rRoTz2g111(O``pv)UjkT7k_z;%loZlDYGvg@ec(h2co zX8Y9_QHDx6ZyRrSg%a_M$l^XKjd?xH(dxOR7X-5y1$={+baaLbG1LVhJ0X7<>@U=)E_DDGd@knv zQNY$ffo@~rB!>XGU*b<<`vPQXOs=C7CvM3d+rNTa5Gk7bxK>oN=iz^ znF+vT+afR|0ENKDD$tYbbUd(@hTt3J9%guGhRZ0Cyt!MQliT%EA;Bkxh)T73GdYZOKhH!6Q zRh6~hU_QMvE#OU4y&BH2JmwF%Nxu8+Peu%Rmm5KB7H z_mGJovtH$dJh0jzF4DZd|CvjWdD8Q!tgz;A@`uyUZ|XZ+6pagG$sc!EKGH>>o7;{Y zrRqxi1EYO$<_(uHzV&G&guu5Z$MiVkx&r6faBU==Bwu?w9!ksfPVpNovmkJU!@}#! z2!LHWip!~^@;!k_?IaK8$INHq78f$rkHitws$Jjha|&`y9x3HnB#n0eeX{TbLI<|v zj=|ByWg3i8D^3f~9|EB^j zFXC0wgON;!_j+vPre#5|m@rS{v6jlkMQYr-{j+8#i!YsI&%T-%##=bGh4)#s{@`(o za8V4eC=qQy|9lJm+duE&mDY1j@pc<&8b6*sqR?O8SP!jzZCOVkW8;3G0_K)7!=l4G z#hjfbO&uTR#}~L|9%$BN*|!XCPIvd`bX6AdZ#pX9FzlxD(p^T-Xr*Us8U0)oYbOsp zWJ2w+P1Y>vJwM&q--RdwPV`;@nQ83}3M5FB}u;%ds z`A1zP?Q}C6b`Y$j?kdiB3AxZqx|eQ|G?rJ7sY-p_#ksv0n#H!9Id;Eg-5Yz|XQuz_ zv_63yp1$s$I(hP!CVbDfrCvbkw0t~OcDq7zEadd)$UJ*OMWg+NhXcU!v^4Yg{7}}t zPEdN%?jkLq^l?TA{IK)t7tl`8!O)J$(KeUo!1&nW?A;PanaQd4D-q6n*jqXeZNtS*_C`{-SqoODeZ=z zmDF42a_>!ugKLa{P^BgRPmFwWCjg)8O$tSSQ>j7gF=CeXjl3LgCUOM||0XNA1i>rv z-GVv?=f<%hl5&cJVBzWu)Z`>bnTN5L8`G|5AaQQw$oI`(019qU{M`bOGB02|izth}-Iw5QPczs7D)jrQTM3lVDl~;8 zINrzLo|mm}hILcQlBcx}iXRzwJh8N9a>;-40y1Hh36PsbQB}LU)^rT1v-h#nW4eb@ zlrf-pTMlad5mo-)*|W4FAD=pr5S)7;oqrOiTVxEK&kCCt(~14c?vSMGU=RO&Ix`0t z{RBp3BS34;U8!0q_0jSgVQr)hpf`$L?6Rlw!|zqqm1k7mO)IBIpv}`p z7igjD&ncp8sFk7x#RmUYOxuq(b>Z-sIgB^i}6$=d?IO@Sj*#24ZM*$Cg@d9kZy5!64iN0~AzfMcxM5B12pI24Ku znjX6PcDi!R3%7^!kzKVo#~nEjkWN3{9Lx{%Q(z$WM`Y8c#+c%i^3b8rx9xcvZZEvW z$k;e~+g{-~%c=d#u?}hkMlNDx%oAsC5(FRN^+8E}3pxa6)%1_%x+^ZzjLP$bIM6eX zNw=eiyqIz5%nM5hwt^rN9Qjl8833k+=%t-l>WZV}&!+tg7d2LXbU6kR1P!1Z*y(;m z04Abp^=7t_Iolc&+EZKYOM$61`cdjTdxNNc6m3=GNe%bYzWMLhN!6v=%$8P8&yQGx zJQ3^jgb*#DuDXdFgG>S{XGctT@09VKe0C8`Bn@wF83*+!+&I>!clK&>f@rFcf_@W5!|MbQZc8r78Nf{f9-aavZ)qg(`=YPJh*}kP&*?z9sG2LLOVfy>iGr{(A zF%pVP;~3Zm9=HEs4Lsd0I-BoiljWgFxN8FXIIgMSgsDMJ1pI3|sn5I>n%>u6VoepRa8FJw)A^BSPU9^N*^ zezOuV)(DYwqk{~%pLL=qQu{~nlAo&m&kGONztLmzgjd7@SY%fisrM}xw_KDq%gUt+ z(b}aNt-#gywj+j{f5tV|I2HpmYmhX^bCd1Tx&ir*I&~D4Wn@QoSo=Vv=US4;+{;P) zY0Ks}g~TOx9V+h@kt>P$y+NsHrO|Z)R+GN&Y-`Zk^%-&J!>IG^m#+W%y3hWjl5o8M zdBGr;!+vVeyspT?va0ZnTDI2VXpIosr_knb@B_XzT>g(QH7av?gOIj)>#5vb_oD9a z^Y9hZ#(NY+Ah*)=@W8=O%N?B#TKfB%p@&7sFiNtdf1#0TJ#kAdw2e{OxWiyl$4o4c z$nbXO7B6WVzl1`f%UOK&o}Xwm&Tfc^(a+Tl8f`H!e5d=GJQx%MY@J7#OqDIKb7g2I>hZ%Aq=N(K8KFUsVHP*862{+NBjvN%cJ zKmecawV1cC1owB;7#{Ho(A~d`#*KqmuEPrk@cB(TUn=NbpCo67^fXv;S%UA^6@~p9 zn0M{%cPZvB?J&LH95A6NZ_cBlPV=Oxig zLZsd;#1HX&?lYgY=m~(?2vQf1pYKkU?M-yWxmoKL40&97&=BW`{vjClrsjj?gHQ=^ z&%oof9)hVw4f%g`$+Wha|qdYj5K^tXCI)9?DTqD%|*ydU@DhZrTo#-X=1hb!OY z?$cY7UWQSseP_z-oE_6sCdr0PhG@t*1tRyTL5xAnqyIuOLtqwBZvyx!2iNz=@woX7 zKOc^WN`=xgxneeHE4Of*sw-OX?Uwpk%Vu!dc_8{*y@~-%c#iiB0d_Mc&EqR zZdpY>zW~r6;|L|D={`jdKJM>T4bj~6L86fx8QFSF@aHeLXq!Wg;D#it1nNPO4bZfW z#FyE(^Hj7sM-fbn@{xQ+7fZ}cg4tsI>gxq#t@(IKu$~x!RwUi-S7UOcPhA2gn++6) z<&@I>g+V8xVogkHbXs}s2s{PGKvP>B?niR5`0A`gqL0HJwx>V!5GCQs@uN2>o%HRUsX|`C$79eLTKTg zxLidi4bLOuSG8QRVVZti$xJ5Ir%olm87&SC$VXb;Ysu(UP_`D?n1iwabzlnDq=j(79^jC=iX-G;5VN6#wb104K=MiykkRY#+}5buu%%9kEKblZ zrZJlw)f7u4H_3};ofRrfEfaZU+hqd=e@j^n$}jlu*)>RX;V7xQU`g#e9&B{cF5G&LZzV8|CBgt7brUAXY^{g8lY%&2pV%`T7k1+At%ARg)Oe@clZ5fN zV?s)-gQke6+wFhFo(Kqcj&E!#)R`&mh2eS>zX!HKE~*{fg{Z6PG6m6)kE}AuFrW4#C?)9ZBU!FTo&~@h|8g~F zG0NLuj+)&dFA(LCZhxAZtjjrCxV)C6XOYgI&RPNI=?{c)fhN&@hqj@!D^XYF0S)Ln zF2&BLs8`&@i-NT{_iY|MM!?2Pg`P!1v^G%=@2R3gJc;;k(0`sH$32t^u!A5X=G4OQCIV+IsDwAx{<3b$I!fX8 zp@SK!1VgH@q{Az`Uax3Yf2flh*YZnf>92c7hHTbk7N9XX_@`E#V}*yMl=9c>usS1= zP$U`S*I<@WJo_vT)7W4QQO_6F z8faC8Bx@n?g`Vdc!c-Xx{`(mE5A@#qPgQ@vss?4TpWN0o>#mKJnXF@L^I~*&tg#*~ z?b1eQH#zX%DA>baGhkGe`__{Evl!TX>Br8=)D~B+Rg>2lNI>dWnu_1P^B}Q|1Xg+w zRS-wv+mG~`hCX?XwF0u3$^=U#`viA;#Xc2c1DH4Jydq6iW_6a7ryCkknS7uT9jG;P2LxagQhi!laX z3dqsQv-fDa(FV$DNRP88>Ol?U?lthe6Im6QrtPgqUS4$f$PhJP=Tqa+q;uI`ZvSon zVa(-VhfREt!@+B#jLv>(oajw9!Y|hmYfYmwOZabA6SaEw2nIhK`$jHWCjtsme=`1&zDK& z=*e!Y_xPKA`;qDKTi{WxyzwPU8H>eadQZq_CX8QHR3!5yx0qH&rJS9;C9f3nZjtBc$0JV3DG7(JxKPk5W(GSEvzn;m0! zo#xs;WMZQySL1}pHgeBasg{dKk5>E}wliQNK(<0)@6;Bc^uZsk!B1mV@#Opa<&lXt z%(;p59a+b_yHw{jHq0%02ne`V*vFu?{r2AnWXFBw%&h@M3vaE+cn^v0^)vN1?Kz)1 zL|FU3h=fqD(P1O~P(zPR;P$OfS(7U6{MLM$C)jnHa_F}#e;2UY2kAsL)GTqv^^*H8 z-@+F%n~5MzrUkm2>9+Ot{#Cur_^u=4)UB_Q)Cj%pjO7Vvp5F#_PUizi4C{paiZg@U z-#_4Ny;2qQnTWbpuRY}O#EnpXu<&H zK|b#vM15Gm%jR8NfxTN%i#Jv(^-U z)~XQ7 zVuN*xN>b!5lJO_42{2C%vL@k4$2V>DL==SD$|D?Ml9sA!NK+}H1MLzj95i`TX2AXA zw72@>BS0$JnsGd2QcWGDG>9qGruHmz02ckgxnkfxef1CL*H~z>ZLiahbBIH*t~GVE z6){O-Ve)K#lroiEO^s-*|J5M;PD4VljVa3-8~%%cRVIAvcjn>vA3v|Rq!y!?-)qfX zxOjDp*v`gdPSDj59}}wn0hRwg14b9#+#E8ru!d0Y$pVh+R)N_(rz^!K-(bDP>QL1m*uc z#aop&D_1ShUaxx|fgy~4%dh{j;modGF;Ru2&6^|S0`_?F5uKFZ0AnS!ja->rz1M-Lk?IAz=9Un%*JMX+m`Cr;i|_!>TTX zg`54bVO_gUAgNf$t`;q5DwyseEZ<+BL(1ky@)Nw2v)F4*JGv~rpbML`Hx%asZTgZu z#0K%#XjxUicmGS)6~`~z<{U|lLU~g3tjNaxYi(9F71%))T0Di=xohgmk6|C!FP7Yv zp=cN12J-)wM_0$k;Ht@^EyhQwIYyc}p+7%R`iH#OfYECbgzskN(nmOKPs^#Ak~ zilp^7C1PAVMUTt@5d&5jQfYs=K;MfoD?G%o^|yecLiu8+r1sA(2W;(AV~w9O3{y2f z8THOttp(NNqoGPX{!H&+o8+I?g?VGA)Fc6udl9=2;1=FB`jR zJU=?3)=vs}en`2G;RRmEwTD{Uu?0vhYh5}ep9=1KMIGY{ZgM45#luBWQCyr$4m6Ua zPyDV3|29N<@I3X8qHJ^yhUQg=gs43y2g%1n>vDcX#)RCVoYw3l)?=iSgt_g_9pQd~ zp@Q%N8O-NAYFX#Qe!Zwe6P799{Mv!th@|hyI6BS^i|tp?y{0caXVUPGb9S0Vm$<`1 zjtQ$aV&k&`Q7@=q=1X5Qv>&fW^d0?qO(vCS{xA(#+H7zRLyOz>Wn>b*#xm n*Aliz+N m.type === kind); + const [selectedModel, setSelectedModel] = useState(kindModels[0]?.id ?? ""); + const [input, setInput] = useState(exConfig.defaultInput); const [apiKey, setApiKey] = useState(""); const [useTunnel, setUseTunnel] = useState(false); @@ -848,9 +852,10 @@ function GenericExampleCard({ providerId, kind }) { const endpoint = useTunnel ? tunnelEndpoint : localEndpoint; const apiPath = kindConfig.endpoint.path; + const modelFull = selectedModel ? `${providerAlias}/${selectedModel}` : ""; const requestBody = { - model: `${providerAlias}/model-name`, + model: modelFull, [exConfig.bodyKey]: input, ...exConfig.extraBody, }; @@ -861,7 +866,7 @@ function GenericExampleCard({ providerId, kind }) { -d '${JSON.stringify(requestBody)}'`; const handleRun = async () => { - if (!input.trim()) return; + if (!input.trim() || !modelFull) return; setRunning(true); setError(""); setResult(null); @@ -869,7 +874,7 @@ function GenericExampleCard({ providerId, kind }) { try { const headers = { "Content-Type": "application/json" }; if (apiKey) headers["Authorization"] = `Bearer ${apiKey}`; - const body = { ...requestBody, model: `${providerAlias}/model-name` }; + const body = { ...requestBody, model: modelFull }; const res = await fetch(`/api${apiPath}`, { method: kindConfig.endpoint.method, headers, @@ -892,6 +897,21 @@ function GenericExampleCard({ providerId, kind }) {

Example

+ {/* Model selector - only show if models available */} + {kindModels.length > 0 && ( + + + + )} + {/* Endpoint */}
@@ -953,11 +973,11 @@ function GenericExampleCard({ providerId, kind }) { {copiedCurl ? "check" : "content_copy"} {copiedCurl ? "Copied" : "Copy"} -
diff --git a/src/app/(dashboard)/dashboard/providers/components/ModelsCard.js b/src/app/(dashboard)/dashboard/providers/components/ModelsCard.js index 5cb32e6e..a9f66d2d 100644 --- a/src/app/(dashboard)/dashboard/providers/components/ModelsCard.js +++ b/src/app/(dashboard)/dashboard/providers/components/ModelsCard.js @@ -111,6 +111,7 @@ AddCustomModelModal.propTypes = { export default function ModelsCard({ providerId, kindFilter }) { const { copied, copy } = useCopyToClipboard(); const [modelAliases, setModelAliases] = useState({}); + const [customModels, setCustomModels] = useState([]); const [modelTestResults, setModelTestResults] = useState({}); const [testingModelId, setTestingModelId] = useState(null); const [testError, setTestError] = useState(""); @@ -118,17 +119,21 @@ export default function ModelsCard({ providerId, kindFilter }) { const [connections, setConnections] = useState([]); const providerAlias = getProviderAlias(providerId); + const effectiveType = kindFilter || "llm"; const fetchData = useCallback(async () => { try { - const [aliasRes, connRes] = await Promise.all([ + const [aliasRes, connRes, customRes] = await Promise.all([ fetch("/api/models/alias"), fetch("/api/providers", { cache: "no-store" }), + fetch("/api/models/custom", { cache: "no-store" }), ]); const aliasData = await aliasRes.json(); const connData = await connRes.json(); + const customData = await customRes.json(); if (aliasRes.ok) setModelAliases(aliasData.aliases || {}); if (connRes.ok) setConnections((connData.connections || []).filter((c) => c.provider === providerId)); + if (customRes.ok) setCustomModels(customData.models || []); } catch (e) { console.log("ModelsCard fetch error:", e); } }, [providerId]); @@ -153,6 +158,25 @@ export default function ModelsCard({ providerId, kindFilter }) { } catch (e) { console.log("delete alias error:", e); } }; + const handleAddCustomModel = async (modelId) => { + try { + const res = await fetch("/api/models/custom", { + method: "POST", + headers: { "Content-Type": "application/json" }, + body: JSON.stringify({ providerAlias, id: modelId, type: effectiveType }), + }); + if (res.ok) await fetchData(); + } catch (e) { console.log("add custom model error:", e); } + }; + + const handleDeleteCustomModel = async (modelId) => { + try { + const params = new URLSearchParams({ providerAlias, id: modelId, type: effectiveType }); + const res = await fetch(`/api/models/custom?${params}`, { method: "DELETE" }); + if (res.ok) await fetchData(); + } catch (e) { console.log("delete custom model error:", e); } + }; + const handleTestModel = async (modelId) => { if (testingModelId) return; setTestingModelId(modelId); @@ -171,28 +195,23 @@ export default function ModelsCard({ providerId, kindFilter }) { } finally { setTestingModelId(null); } }; - // Get models — filter by kindFilter if provided - const allModels = getModelsByProviderId(providerId); - const displayModels = kindFilter - ? allModels.filter((m) => { + // Built-in models — filter by kindFilter if provided + const allBuiltIn = getModelsByProviderId(providerId); + const builtInModels = kindFilter + ? allBuiltIn.filter((m) => { if (m.kinds) return m.kinds.includes(kindFilter); - if (m.type) return m.type === kindFilter; - return kindFilter === "llm"; + return (m.type || "llm") === kindFilter; }) - : allModels; + : allBuiltIn; - // Custom models added via alias - const customModels = Object.entries(modelAliases) - .filter(([alias, fullModel]) => { - const prefix = `${providerAlias}/`; - if (!fullModel.startsWith(prefix)) return false; - const modelId = fullModel.slice(prefix.length); - return !displayModels.some((m) => m.id === modelId) && alias === modelId; - }) - .map(([alias, fullModel]) => ({ - id: fullModel.slice(`${providerAlias}/`.length), - alias, - })); + // Custom models for this provider + kind, dedupe vs built-in + const myCustomModels = customModels.filter( + (m) => m.providerAlias === providerAlias + && (m.type || "llm") === effectiveType + && !builtInModels.some((b) => b.id === m.id) + ); + + const displayModels = builtInModels; return ( <> @@ -224,16 +243,15 @@ export default function ModelsCard({ providerId, kindFilter }) { ); })} - {customModels.map((model) => ( + {myCustomModels.map((model) => ( {}} - onDeleteAlias={() => handleDeleteAlias(model.alias)} + onDeleteAlias={() => handleDeleteCustomModel(model.id)} testStatus={modelTestResults[model.id]} onTest={connections.length > 0 ? () => handleTestModel(model.id) : undefined} isTesting={testingModelId === model.id} @@ -254,7 +272,7 @@ export default function ModelsCard({ providerId, kindFilter }) { { - await handleSetAlias(modelId, modelId); + await handleAddCustomModel(modelId); setShowAddCustomModel(false); }} onClose={() => setShowAddCustomModel(false)} diff --git a/src/app/api/models/custom/route.js b/src/app/api/models/custom/route.js new file mode 100644 index 00000000..76e14c96 --- /dev/null +++ b/src/app/api/models/custom/route.js @@ -0,0 +1,48 @@ +import { NextResponse } from "next/server"; +import { getCustomModels, addCustomModel, deleteCustomModel } from "@/models"; + +export const dynamic = "force-dynamic"; + +// GET /api/models/custom - List all custom models +export async function GET() { + try { + const models = await getCustomModels(); + return NextResponse.json({ models }); + } catch (error) { + console.log("Error fetching custom models:", error); + return NextResponse.json({ error: "Failed to fetch custom models" }, { status: 500 }); + } +} + +// POST /api/models/custom - Add custom model +export async function POST(request) { + try { + const { providerAlias, id, type, name } = await request.json(); + if (!providerAlias || !id) { + return NextResponse.json({ error: "providerAlias and id required" }, { status: 400 }); + } + const added = await addCustomModel({ providerAlias, id, type: type || "llm", name }); + return NextResponse.json({ success: true, added }); + } catch (error) { + console.log("Error adding custom model:", error); + return NextResponse.json({ error: "Failed to add custom model" }, { status: 500 }); + } +} + +// DELETE /api/models/custom?providerAlias=xxx&id=yyy&type=zzz +export async function DELETE(request) { + try { + const { searchParams } = new URL(request.url); + const providerAlias = searchParams.get("providerAlias"); + const id = searchParams.get("id"); + const type = searchParams.get("type") || "llm"; + if (!providerAlias || !id) { + return NextResponse.json({ error: "providerAlias and id required" }, { status: 400 }); + } + await deleteCustomModel({ providerAlias, id, type }); + return NextResponse.json({ success: true }); + } catch (error) { + console.log("Error deleting custom model:", error); + return NextResponse.json({ error: "Failed to delete custom model" }, { status: 500 }); + } +} diff --git a/src/app/api/v1/images/generations/route.js b/src/app/api/v1/images/generations/route.js new file mode 100644 index 00000000..302a718b --- /dev/null +++ b/src/app/api/v1/images/generations/route.js @@ -0,0 +1,16 @@ +import { handleImageGeneration } from "@/sse/handlers/imageGeneration.js"; + +export async function OPTIONS() { + return new Response(null, { + headers: { + "Access-Control-Allow-Origin": "*", + "Access-Control-Allow-Methods": "POST, OPTIONS", + "Access-Control-Allow-Headers": "*", + }, + }); +} + +/** POST /v1/images/generations - OpenAI-compatible image generation endpoint */ +export async function POST(request) { + return await handleImageGeneration(request); +} diff --git a/src/lib/localDb.js b/src/lib/localDb.js index ced9b3c8..f2c2cafd 100644 --- a/src/lib/localDb.js +++ b/src/lib/localDb.js @@ -44,6 +44,7 @@ function cloneDefaultData() { providerNodes: [], proxyPools: [], modelAliases: {}, + customModels: [], mitmAlias: {}, combos: [], apiKeys: [], @@ -515,6 +516,33 @@ export async function deleteModelAlias(alias) { await safeWrite(db); } +// Custom models — user-added models with explicit type (llm/image/tts/embedding/...) +export async function getCustomModels() { + const db = await getDb(); + return db.data.customModels || []; +} + +export async function addCustomModel({ providerAlias, id, type = "llm", name }) { + const db = await getDb(); + if (!db.data.customModels) db.data.customModels = []; + const exists = db.data.customModels.some( + (m) => m.providerAlias === providerAlias && m.id === id && (m.type || "llm") === type + ); + if (exists) return false; + db.data.customModels.push({ providerAlias, id, type, name: name || id }); + await safeWrite(db); + return true; +} + +export async function deleteCustomModel({ providerAlias, id, type = "llm" }) { + const db = await getDb(); + if (!db.data.customModels) return; + db.data.customModels = db.data.customModels.filter( + (m) => !(m.providerAlias === providerAlias && m.id === id && (m.type || "llm") === type) + ); + await safeWrite(db); +} + export async function getMitmAlias(toolName) { const db = await getDb(); const all = db.data.mitmAlias || {}; diff --git a/src/models/index.js b/src/models/index.js index e61129fe..99444f62 100644 --- a/src/models/index.js +++ b/src/models/index.js @@ -25,6 +25,9 @@ export { getModelAliases, setModelAlias, deleteModelAlias, + getCustomModels, + addCustomModel, + deleteCustomModel, getMitmAlias, setMitmAliasAll, getApiKeys, diff --git a/src/shared/components/Sidebar.js b/src/shared/components/Sidebar.js index 50ab12ed..eee07906 100644 --- a/src/shared/components/Sidebar.js +++ b/src/shared/components/Sidebar.js @@ -12,7 +12,7 @@ import Button from "./Button"; import { ConfirmModal } from "./Modal"; // const VISIBLE_MEDIA_KINDS = ["embedding", "image", "imageToText", "tts", "stt", "webSearch", "webFetch", "video", "music"]; -const VISIBLE_MEDIA_KINDS = ["embedding", "tts"]; +const VISIBLE_MEDIA_KINDS = ["embedding", "image", "tts"]; const navItems = [ { href: "/dashboard/endpoint", label: "Endpoint", icon: "api" }, diff --git a/src/shared/constants/providers.js b/src/shared/constants/providers.js index 6fc007cb..a024080e 100644 --- a/src/shared/constants/providers.js +++ b/src/shared/constants/providers.js @@ -9,7 +9,7 @@ export const FREE_PROVIDERS = { // codebuddy: { id: "codebuddy", alias: "cb", name: "CodeBuddy", icon: "smart_toy", color: "#006EFF" }, // qoder: { id: "qoder", alias: "qd", name: "Qoder AI", icon: "water_drop", color: "#EC4899" }, iflow: { id: "iflow", alias: "if", name: "iFlow AI", icon: "water_drop", color: "#6366F1" }, - opencode: { id: "opencode", alias: "oc", name: "OpenCode", icon: "terminal", color: "#E87040", textIcon: "OC", noAuth: true, passthroughModels: true, modelsFetcher: { url: "https://opencode.ai/zen/v1/models", type: "opencode-free" } }, + opencode: { id: "opencode", alias: "oc", name: "OpenCode Free", icon: "terminal", color: "#E87040", textIcon: "OC", noAuth: true, passthroughModels: true, modelsFetcher: { url: "https://opencode.ai/zen/v1/models", type: "opencode-free" } }, }; // Free Tier Providers (has free access but may require account/API key) @@ -61,6 +61,7 @@ export const APIKEY_PROVIDERS = { "alicode-intl": { id: "alicode-intl", alias: "alicode-intl", name: "Alibaba Intl", icon: "cloud", color: "#FF6A00", textIcon: "ALi" }, openai: { id: "openai", alias: "openai", name: "OpenAI", icon: "auto_awesome", color: "#10A37F", textIcon: "OA", website: "https://platform.openai.com", serviceKinds: ["llm", "embedding", "tts", "image", "imageToText", "webSearch"], thinkingConfig: THINKING_CONFIG.effort }, anthropic: { id: "anthropic", alias: "anthropic", name: "Anthropic", icon: "smart_toy", color: "#D97757", textIcon: "AN", website: "https://console.anthropic.com", serviceKinds: ["llm", "imageToText"] }, + "opencode-go": { id: "opencode-go", alias: "ocg", name: "OpenCode Go", icon: "terminal", color: "#E87040", textIcon: "OC", website: "https://opencode.ai/auth", notice: { text: "OpenCode Go subscription: $5/mo (then $10/mo). Access to Kimi, GLM, Qwen, MiMo, MiniMax models.", apiKeyUrl: "https://opencode.ai/auth" } }, deepseek: { id: "deepseek", alias: "ds", name: "DeepSeek", icon: "bolt", color: "#4D6BFE", textIcon: "DS", website: "https://deepseek.com" }, diff --git a/src/sse/handlers/imageGeneration.js b/src/sse/handlers/imageGeneration.js new file mode 100644 index 00000000..47911958 --- /dev/null +++ b/src/sse/handlers/imageGeneration.js @@ -0,0 +1,152 @@ +import { + getProviderCredentials, + markAccountUnavailable, + clearAccountError, + extractApiKey, + isValidApiKey, +} from "../services/auth.js"; +import { getSettings } from "@/lib/localDb"; +import { getModelInfo } from "../services/model.js"; +import { handleImageGenerationCore } from "open-sse/handlers/imageGenerationCore.js"; +import { errorResponse, unavailableResponse } from "open-sse/utils/error.js"; +import { HTTP_STATUS } from "open-sse/config/runtimeConfig.js"; +import * as log from "../utils/logger.js"; +import { updateProviderCredentials, checkAndRefreshToken } from "../services/tokenRefresh.js"; + +// Providers that don't require credentials (noAuth) +const NO_AUTH_PROVIDERS = new Set(["sdwebui", "comfyui"]); + +/** + * Handle image generation request + * @param {Request} request + */ +export async function handleImageGeneration(request) { + let body; + try { + body = await request.json(); + } catch { + log.warn("IMAGE", "Invalid JSON body"); + return errorResponse(HTTP_STATUS.BAD_REQUEST, "Invalid JSON body"); + } + + const url = new URL(request.url); + const modelStr = body.model; + + log.request("POST", `${url.pathname} | ${modelStr}`); + + const apiKey = extractApiKey(request); + if (apiKey) { + log.debug("AUTH", `API Key: ${log.maskKey(apiKey)}`); + } else { + log.debug("AUTH", "No API key provided (local mode)"); + } + + const settings = await getSettings(); + if (settings.requireApiKey) { + if (!apiKey) { + log.warn("AUTH", "Missing API key (requireApiKey=true)"); + return errorResponse(HTTP_STATUS.UNAUTHORIZED, "Missing API key"); + } + const valid = await isValidApiKey(apiKey); + if (!valid) { + log.warn("AUTH", "Invalid API key (requireApiKey=true)"); + return errorResponse(HTTP_STATUS.UNAUTHORIZED, "Invalid API key"); + } + } + + if (!modelStr) { + log.warn("IMAGE", "Missing model"); + return errorResponse(HTTP_STATUS.BAD_REQUEST, "Missing model"); + } + + if (!body.prompt) { + log.warn("IMAGE", "Missing prompt"); + return errorResponse(HTTP_STATUS.BAD_REQUEST, "Missing required field: prompt"); + } + + const modelInfo = await getModelInfo(modelStr); + if (!modelInfo.provider) { + log.warn("IMAGE", "Invalid model format", { model: modelStr }); + return errorResponse(HTTP_STATUS.BAD_REQUEST, "Invalid model format"); + } + + const { provider, model } = modelInfo; + + if (modelStr !== `${provider}/${model}`) { + log.info("ROUTING", `${modelStr} → ${provider}/${model}`); + } else { + log.info("ROUTING", `Provider: ${provider}, Model: ${model}`); + } + + // noAuth providers — no credential needed + if (NO_AUTH_PROVIDERS.has(provider)) { + const result = await handleImageGenerationCore({ + body, + modelInfo: { provider, model }, + credentials: null, + log, + }); + if (result.success) return result.response; + return errorResponse(result.status || HTTP_STATUS.BAD_GATEWAY, result.error || "Image generation failed"); + } + + // Credentialed providers — fallback loop + const excludeConnectionIds = new Set(); + let lastError = null; + let lastStatus = null; + + while (true) { + const credentials = await getProviderCredentials(provider, excludeConnectionIds, model); + + if (!credentials || credentials.allRateLimited) { + if (credentials?.allRateLimited) { + const errorMsg = lastError || credentials.lastError || "Unavailable"; + const status = lastStatus || Number(credentials.lastErrorCode) || HTTP_STATUS.SERVICE_UNAVAILABLE; + log.warn("IMAGE", `[${provider}/${model}] ${errorMsg} (${credentials.retryAfterHuman})`); + return unavailableResponse(status, `[${provider}/${model}] ${errorMsg}`, credentials.retryAfter, credentials.retryAfterHuman); + } + if (excludeConnectionIds.size === 0) { + log.error("AUTH", `No credentials for provider: ${provider}`); + return errorResponse(HTTP_STATUS.BAD_REQUEST, `No credentials for provider: ${provider}`); + } + log.warn("IMAGE", "No more accounts available", { provider }); + return errorResponse(lastStatus || HTTP_STATUS.SERVICE_UNAVAILABLE, lastError || "All accounts unavailable"); + } + + log.info("AUTH", `\x1b[32mUsing ${provider} account: ${credentials.connectionName}\x1b[0m`); + + const refreshedCredentials = await checkAndRefreshToken(provider, credentials); + + const result = await handleImageGenerationCore({ + body, + modelInfo: { provider, model }, + credentials: refreshedCredentials, + log, + onCredentialsRefreshed: async (newCreds) => { + await updateProviderCredentials(credentials.connectionId, { + accessToken: newCreds.accessToken, + refreshToken: newCreds.refreshToken, + providerSpecificData: newCreds.providerSpecificData, + testStatus: "active" + }); + }, + onRequestSuccess: async () => { + await clearAccountError(credentials.connectionId, credentials, model); + } + }); + + if (result.success) return result.response; + + const { shouldFallback } = await markAccountUnavailable(credentials.connectionId, result.status, result.error, provider, model); + + if (shouldFallback) { + log.warn("AUTH", `Account ${credentials.connectionName} unavailable (${result.status}), trying fallback`); + excludeConnectionIds.add(credentials.connectionId); + lastError = result.error; + lastStatus = result.status; + continue; + } + + return result.response; + } +} diff --git a/tests/unit/image-generation.test.js b/tests/unit/image-generation.test.js new file mode 100644 index 00000000..fe40ac03 --- /dev/null +++ b/tests/unit/image-generation.test.js @@ -0,0 +1,320 @@ +/** + * Unit tests for image generation handler + * + * Covers: + * - OpenAI-compatible format (openai, minimax, openrouter) + * - Gemini format (generateContent API) + * - Provider-specific formats (nanobanana, sdwebui) + * - Response normalization to OpenAI format + * - Error handling (missing prompt, invalid model) + */ + +import { describe, it, expect, vi, beforeEach, afterEach } from "vitest"; +import { handleImageGenerationCore } from "../../open-sse/handlers/imageGenerationCore.js"; + +const originalFetch = global.fetch; + +describe("handleImageGenerationCore", () => { + beforeEach(() => { + global.fetch = vi.fn(); + }); + + afterEach(() => { + global.fetch = originalFetch; + }); + + it("validates required prompt field", async () => { + const result = await handleImageGenerationCore({ + body: { model: "openai/dall-e-3" }, + modelInfo: { provider: "openai", model: "dall-e-3" }, + credentials: { apiKey: "test-key" }, + log: null, + }); + + expect(result.success).toBe(false); + expect(result.status).toBe(400); + expect(result.error).toContain("Missing required field: prompt"); + }); + + it("rejects unsupported provider", async () => { + const result = await handleImageGenerationCore({ + body: { prompt: "test" }, + modelInfo: { provider: "unknown-provider", model: "test" }, + credentials: null, + log: null, + }); + + expect(result.success).toBe(false); + expect(result.status).toBe(400); + expect(result.error).toContain("does not support image generation"); + }); + + it("generates image with OpenAI format", async () => { + global.fetch.mockResolvedValueOnce( + new Response( + JSON.stringify({ + created: 1234567890, + data: [{ url: "https://example.com/image.png" }], + }), + { status: 200, headers: { "Content-Type": "application/json" } } + ) + ); + + const result = await handleImageGenerationCore({ + body: { prompt: "A cute cat", n: 1, size: "1024x1024" }, + modelInfo: { provider: "openai", model: "dall-e-3" }, + credentials: { apiKey: "test-key" }, + log: null, + }); + + expect(result.success).toBe(true); + expect(global.fetch).toHaveBeenCalledWith( + "https://api.openai.com/v1/images/generations", + expect.objectContaining({ + method: "POST", + headers: expect.objectContaining({ + "Content-Type": "application/json", + Authorization: "Bearer test-key", + }), + body: expect.stringContaining('"prompt":"A cute cat"'), + }) + ); + + const responseBody = await result.response.json(); + expect(responseBody.data).toHaveLength(1); + expect(responseBody.data[0].url).toBe("https://example.com/image.png"); + }); + + it("generates image with Gemini format", async () => { + global.fetch.mockResolvedValueOnce( + new Response( + JSON.stringify({ + candidates: [ + { + content: { + parts: [ + { text: "Generated image" }, + { inlineData: { data: "base64imagedata" } }, + ], + }, + }, + ], + }), + { status: 200, headers: { "Content-Type": "application/json" } } + ) + ); + + const result = await handleImageGenerationCore({ + body: { prompt: "A sunset" }, + modelInfo: { provider: "gemini", model: "gemini-image-preview" }, + credentials: { apiKey: "test-key" }, + log: null, + }); + + expect(result.success).toBe(true); + expect(global.fetch).toHaveBeenCalledWith( + expect.stringContaining("generativelanguage.googleapis.com"), + expect.objectContaining({ + method: "POST", + body: expect.stringContaining('"responseModalities":["TEXT","IMAGE"]'), + }) + ); + + const responseBody = await result.response.json(); + expect(responseBody.data).toHaveLength(1); + expect(responseBody.data[0].b64_json).toBe("base64imagedata"); + }); + + it("generates image with Minimax format", async () => { + global.fetch.mockResolvedValueOnce( + new Response( + JSON.stringify({ + created: 1234567890, + data: [{ url: "https://example.com/minimax.png" }], + }), + { status: 200, headers: { "Content-Type": "application/json" } } + ) + ); + + const result = await handleImageGenerationCore({ + body: { prompt: "A mountain", size: "1024x1024" }, + modelInfo: { provider: "minimax", model: "minimax-image-01" }, + credentials: { apiKey: "test-key" }, + log: null, + }); + + expect(result.success).toBe(true); + expect(global.fetch).toHaveBeenCalledWith( + "https://api.minimaxi.com/v1/images/generations", + expect.objectContaining({ + method: "POST", + headers: expect.objectContaining({ + Authorization: "Bearer test-key", + }), + }) + ); + }); + + it("generates image with NanoBanana format", async () => { + global.fetch.mockResolvedValueOnce( + new Response( + JSON.stringify({ image: "base64nanobanana" }), + { status: 200, headers: { "Content-Type": "application/json" } } + ) + ); + + const result = await handleImageGenerationCore({ + body: { prompt: "A robot", n: 2, size: "1024x1792" }, + modelInfo: { provider: "nanobanana", model: "nanobanana-flash" }, + credentials: { apiKey: "test-key" }, + log: null, + }); + + expect(result.success).toBe(true); + const fetchCall = global.fetch.mock.calls[0]; + const requestBody = JSON.parse(fetchCall[1].body); + expect(requestBody.type).toBe("TEXTTOIAMGE"); + expect(requestBody.numImages).toBe(2); + expect(requestBody.image_size).toBe("9:16"); + + const responseBody = await result.response.json(); + expect(responseBody.data[0].b64_json).toBe("base64nanobanana"); + }); + + it("generates image with SD WebUI format", async () => { + global.fetch.mockResolvedValueOnce( + new Response( + JSON.stringify({ images: ["base64sdwebui1", "base64sdwebui2"] }), + { status: 200, headers: { "Content-Type": "application/json" } } + ) + ); + + const result = await handleImageGenerationCore({ + body: { prompt: "A forest", size: "768x768", n: 2 }, + modelInfo: { provider: "sdwebui", model: "sdxl-base-1.0" }, + credentials: null, + log: null, + }); + + expect(result.success).toBe(true); + const fetchCall = global.fetch.mock.calls[0]; + const requestBody = JSON.parse(fetchCall[1].body); + expect(requestBody.width).toBe(768); + expect(requestBody.height).toBe(768); + expect(requestBody.batch_size).toBe(2); + + const responseBody = await result.response.json(); + expect(responseBody.data).toHaveLength(2); + }); + + it("handles OpenRouter with HTTP-Referer header", async () => { + global.fetch.mockResolvedValueOnce( + new Response( + JSON.stringify({ + created: 1234567890, + data: [{ url: "https://example.com/or.png" }], + }), + { status: 200, headers: { "Content-Type": "application/json" } } + ) + ); + + const result = await handleImageGenerationCore({ + body: { prompt: "A city" }, + modelInfo: { provider: "openrouter", model: "openai/dall-e-3" }, + credentials: { apiKey: "test-key" }, + log: null, + }); + + expect(result.success).toBe(true); + expect(global.fetch).toHaveBeenCalledWith( + "https://openrouter.ai/api/v1/images/generations", + expect.objectContaining({ + headers: expect.objectContaining({ + "HTTP-Referer": "https://endpoint-proxy.local", + "X-Title": "Endpoint Proxy", + }), + }) + ); + }); + + it("handles HuggingFace binary response", async () => { + const imageBuffer = new Uint8Array([0x89, 0x50, 0x4e, 0x47]); // PNG header + global.fetch.mockResolvedValueOnce( + new Response(imageBuffer, { + status: 200, + headers: { "Content-Type": "image/png" }, + }) + ); + + const result = await handleImageGenerationCore({ + body: { prompt: "A tree" }, + modelInfo: { provider: "huggingface", model: "black-forest-labs/FLUX.1-schnell" }, + credentials: { apiKey: "test-key" }, + log: null, + }); + + expect(result.success).toBe(true); + const responseBody = await result.response.json(); + expect(responseBody.data[0].b64_json).toBeTruthy(); + }); + + it("handles provider error responses", async () => { + global.fetch.mockResolvedValueOnce( + new Response( + JSON.stringify({ error: { message: "Rate limit exceeded" } }), + { status: 429, headers: { "Content-Type": "application/json" } } + ) + ); + + const result = await handleImageGenerationCore({ + body: { prompt: "test" }, + modelInfo: { provider: "openai", model: "dall-e-3" }, + credentials: { apiKey: "test-key" }, + log: null, + }); + + expect(result.success).toBe(false); + expect(result.status).toBe(429); + expect(result.error).toContain("Rate limit exceeded"); + }); + + it("handles network errors", async () => { + global.fetch.mockRejectedValueOnce(new Error("Network timeout")); + + const result = await handleImageGenerationCore({ + body: { prompt: "test" }, + modelInfo: { provider: "openai", model: "dall-e-3" }, + credentials: { apiKey: "test-key" }, + log: null, + }); + + expect(result.success).toBe(false); + expect(result.status).toBe(502); + expect(result.error).toContain("Network timeout"); + }); + + it("calls onRequestSuccess callback on success", async () => { + global.fetch.mockResolvedValueOnce( + new Response( + JSON.stringify({ + created: 1234567890, + data: [{ url: "https://example.com/success.png" }], + }), + { status: 200, headers: { "Content-Type": "application/json" } } + ) + ); + + const onRequestSuccess = vi.fn(); + + const result = await handleImageGenerationCore({ + body: { prompt: "test" }, + modelInfo: { provider: "openai", model: "dall-e-3" }, + credentials: { apiKey: "test-key" }, + log: null, + onRequestSuccess, + }); + + expect(result.success).toBe(true); + expect(onRequestSuccess).toHaveBeenCalledTimes(1); + }); +}); diff --git a/tests/vitest.config.js b/tests/vitest.config.js index e5cd9ae8..0df209bf 100644 --- a/tests/vitest.config.js +++ b/tests/vitest.config.js @@ -16,6 +16,8 @@ export default defineConfig({ alias: { // Resolve open-sse/* imports to the actual local package "open-sse": resolve(__dirname, "../open-sse"), + // Resolve @/* imports to src directory + "@": resolve(__dirname, "../src"), }, }, });