From 876d837c3301c03eb09fe3cad2692ea3acf47105 Mon Sep 17 00:00:00 2001 From: Martin McKeaveney Date: Fri, 20 Sep 2024 14:58:10 +0100 Subject: [PATCH] adjust instantiation of llm class after static initialiser change --- packages/pro | 2 +- packages/server/src/automations/steps/openai.ts | 3 +-- 2 files changed, 2 insertions(+), 3 deletions(-) diff --git a/packages/pro b/packages/pro index 5929fbb3c6..922431260e 160000 --- a/packages/pro +++ b/packages/pro @@ -1 +1 @@ -Subproject commit 5929fbb3c69b46c525faef4a5a805e0a0015bd8e +Subproject commit 922431260e90d558a1ca55398475412e75088057 diff --git a/packages/server/src/automations/steps/openai.ts b/packages/server/src/automations/steps/openai.ts index 5eff12db47..b1dfa3df5b 100644 --- a/packages/server/src/automations/steps/openai.ts +++ b/packages/server/src/automations/steps/openai.ts @@ -103,8 +103,7 @@ export async function run({ const budibaseAIEnabled = await pro.features.isBudibaseAIEnabled() if (budibaseAIEnabled || customConfigsEnabled) { - const llm = new pro.ai.LargeLanguageModel(inputs.model) - await llm.init() + const llm = await pro.ai.LargeLanguageModel.forCurrentTenant(inputs.model) response = await llm.run(inputs.prompt) } else { // fallback to the default that uses the environment variable for backwards compat