From 7fcff8a628bb1bd414e962f46ac504cbd57f785e Mon Sep 17 00:00:00 2001 From: Martin McKeaveney Date: Wed, 4 Dec 2024 10:15:22 +0000 Subject: [PATCH] rely on llm status rather than initialised variable --- packages/pro | 2 +- packages/server/src/utilities/rowProcessor/utils.ts | 4 ++-- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/packages/pro b/packages/pro index e60f4b1b364..0d7fa31d4c4 160000 --- a/packages/pro +++ b/packages/pro @@ -1 +1 @@ -Subproject commit e60f4b1b364fd49d2bb082f298757f83cb2032f0 +Subproject commit 0d7fa31d4c4019690e2200323421025cdc74b89e diff --git a/packages/server/src/utilities/rowProcessor/utils.ts b/packages/server/src/utilities/rowProcessor/utils.ts index 9dbeb8ebb27..3728401ab82 100644 --- a/packages/server/src/utilities/rowProcessor/utils.ts +++ b/packages/server/src/utilities/rowProcessor/utils.ts @@ -126,8 +126,8 @@ export async function processAIColumns( const numRows = Array.isArray(inputRows) ? inputRows.length : 1 span?.addTags({ table_id: table._id, numRows }) const rows = Array.isArray(inputRows) ? inputRows : [inputRows] - const llm = await pro.ai.LargeLanguageModel.forCurrentTenant("gpt-4o-mini") - if (rows && llm.initialised) { + const llmWrapper = await pro.ai.LargeLanguageModel.forCurrentTenant("gpt-4o-mini") + if (rows && llmWrapper.llm) { // Ensure we have snippet context await context.ensureSnippetContext()