From 63e342b3d52e5de6870f6126e98bb87c13932437 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Jan=20Pokorn=C3=BD?= Date: Thu, 17 Oct 2024 17:30:56 +0200 Subject: [PATCH] fix(code-interpreter): switch to llama --- src/runs/execution/factory.ts | 6 +++--- src/runs/execution/tools/helpers.ts | 2 +- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/src/runs/execution/factory.ts b/src/runs/execution/factory.ts index 5c36e6c..e6185cd 100644 --- a/src/runs/execution/factory.ts +++ b/src/runs/execution/factory.ts @@ -165,7 +165,7 @@ export function createCodeLLM(backend: LLMBackend = LLM_BACKEND) { }); return new IBMvLLM({ client: vllmClient, - modelId: 'ibm/granite-34b-code-instruct', + modelId: 'meta-llama/llama-3-1-70b-instruct', parameters: { method: 'GREEDY', stopping: { include_stop_sequence: false } } }); } @@ -173,7 +173,7 @@ export function createCodeLLM(backend: LLMBackend = LLM_BACKEND) { bamClient ??= new BAMClient({ apiKey: BAM_API_KEY ?? undefined }); return new BAMLLM({ client: bamClient, - modelId: 'ibm/granite-34b-code-instruct', + modelId: 'meta-llama/llama-3-1-70b-instruct', parameters: { decoding_method: 'greedy', include_stop_sequence: false @@ -184,7 +184,7 @@ export function createCodeLLM(backend: LLMBackend = LLM_BACKEND) { if (!WATSONX_API_KEY) throw new Error('Missing WATSONX_API_KEY'); if (!WATSONX_PROJECT_ID) throw new Error('Missing WATSONX_PROJECT_ID'); return new WatsonXLLM({ - modelId: 'ibm/granite-34b-code-instruct', + modelId: 'meta-llama/llama-3-1-70b-instruct', apiKey: WATSONX_API_KEY, projectId: WATSONX_PROJECT_ID, parameters: { diff --git a/src/runs/execution/tools/helpers.ts b/src/runs/execution/tools/helpers.ts index 928e756..62fbc16 100644 --- a/src/runs/execution/tools/helpers.ts +++ b/src/runs/execution/tools/helpers.ts @@ -185,7 +185,7 @@ export async function getTools(run: LoadedRun, context: AgentContext): Promise