From 04be7a66c0a46745e66cbd3212d98a51486bc7b7 Mon Sep 17 00:00:00 2001 From: Riccardo Senica Date: Mon, 19 Jan 2026 20:19:54 +0100 Subject: [PATCH] fix: use lazy loading --- utils/aiGatewayClient.ts | 17 ++++++++++++----- 1 file changed, 12 insertions(+), 5 deletions(-) diff --git a/utils/aiGatewayClient.ts b/utils/aiGatewayClient.ts index 1d250ad..9d439a7 100644 --- a/utils/aiGatewayClient.ts +++ b/utils/aiGatewayClient.ts @@ -1,10 +1,17 @@ import 'dotenv/config'; import OpenAI from 'openai'; -const ovhAI = new OpenAI({ - apiKey: process.env.OVHCLOUD_API_KEY, - baseURL: 'https://oai.endpoints.kepler.ai.cloud.ovh.net/v1' -}); +let ovhAI: OpenAI | null = null; + +function getClient(): OpenAI { + if (!ovhAI) { + ovhAI = new OpenAI({ + apiKey: process.env.OVHCLOUD_API_KEY, + baseURL: 'https://oai.endpoints.kepler.ai.cloud.ovh.net/v1' + }); + } + return ovhAI; +} export interface BaseTool { readonly name: string; @@ -28,7 +35,7 @@ export async function makeRequest( toolDef: T ): Promise> { try { - const completion = await ovhAI.chat.completions.create({ + const completion = await getClient().chat.completions.create({ model: 'Meta-Llama-3_3-70B-Instruct', temperature: 1, max_tokens: 16000,