diff --git a/src/llm/multi-agent/blueberry.ts b/src/llm/multi-agent/blueberry.ts index 57a9635b..aa6defef 100644 --- a/src/llm/multi-agent/blueberry.ts +++ b/src/llm/multi-agent/blueberry.ts @@ -2,7 +2,7 @@ import { BaseLLM } from '#llm/base-llm'; import { GenerateTextOptions, LLM } from '#llm/llm'; import { getLLM } from '#llm/llmFactory'; import { Claude3_5_Sonnet_Vertex } from '#llm/services/anthropic-vertex'; -import { cerebrasLlama3_70b } from '#llm/services/cerebras'; +import { cerebrasLlama3_3_70b } from '#llm/services/cerebras'; import { fireworksLlama3_405B } from '#llm/services/fireworks'; import { GPT4o } from '#llm/services/openai'; import { Gemini_1_5_Flash, Gemini_1_5_Pro } from '#llm/services/vertexai'; @@ -108,7 +108,7 @@ export class Blueberry extends BaseLLM { } } // if (!this.llms) this.llms = [Claude3_5_Sonnet_Vertex(), GPT4o(), Gemini_1_5_Pro(), Claude3_5_Sonnet_Vertex(), fireworksLlama3_405B()]; - let llm = cerebrasLlama3_70b(); + let llm = cerebrasLlama3_3_70b(); // llm = groqLlama3_1_70B(); llm = Gemini_1_5_Flash(); if (!this.llms) this.llms = [llm, llm, llm, llm, llm]; diff --git a/src/llm/services/cerebras.ts b/src/llm/services/cerebras.ts index 9ec23944..3ed20686 100644 --- a/src/llm/services/cerebras.ts +++ b/src/llm/services/cerebras.ts @@ -7,8 +7,7 @@ import { withActiveSpan } from '#o11y/trace'; import { currentUser } from '#user/userService/userContext'; import { RetryableError } from '../../cache/cacheRetry'; import { BaseLLM } from '../base-llm'; -import { GenerateTextOptions, LLM, LlmMessage, combinePrompts } from '../llm'; - +import { GenerateTextOptions, LLM, LlmMessage } from '../llm'; import SystemMessageRequest = CompletionCreateParams.SystemMessageRequest; import AssistantMessageRequest = CompletionCreateParams.AssistantMessageRequest; import UserMessageRequest = CompletionCreateParams.UserMessageRequest; @@ -19,14 +18,13 @@ export const CEREBRAS_SERVICE = 'cerebras'; export function cerebrasLLMRegistry(): Record LLM> { return { 'cerebras:llama3.1-8b': cerebrasLlama3_8b, - 'cerebras:llama3.1-70b': cerebrasLlama3_70b, + 'cerebras:llama3.3-70b': cerebrasLlama3_3_70b, }; } export function cerebrasLlama3_8b(): LLM { return new CerebrasLLM( - 'LLaMA3 8b (Cerebras)', - CEREBRAS_SERVICE, + 'Llama 3.1 8b (Cerebras)', 'llama3.1-8b', 8_192, (input: string) => 0, //(input.length * 0.05) / (1_000_000 * 4), @@ -36,11 +34,10 @@ export function cerebrasLlama3_8b(): LLM { ); } -export function cerebrasLlama3_70b(): LLM { +export function cerebrasLlama3_3_70b(): LLM { return new CerebrasLLM( - 'LLaMA3 70b (Cerebras)', - CEREBRAS_SERVICE, - 'llama3.1-70b', + 'Llama 3.3 70b (Cerebras)', + 'llama-3.3-70b', 8_192, (input: string) => 0, //(input.length * 0.05) / (1_000_000 * 4), (output: string) => 0, //(output.length * 0.08) / (1_000_000 * 4), @@ -57,7 +54,6 @@ export class CerebrasLLM extends BaseLLM { constructor( displayName: string, - service: string, model: string, maxInputTokens: number, /** Needed for Aider when we only have the text size */ @@ -67,15 +63,13 @@ export class CerebrasLLM extends BaseLLM { private costPerMillionInputTokens: number, private costPerMillionOutputTokens: number, ) { - super(displayName, service, model, maxInputTokens, calculateInputCost, calculateOutputCost); + super(displayName, CEREBRAS_SERVICE, model, maxInputTokens, calculateInputCost, calculateOutputCost); } client(): Cerebras { - if (!this._client) { - this._client = new Cerebras({ - apiKey: currentUser().llmConfig.cerebrasKey || process.env.CEREBRAS_API_KEY, - }); - } + this._client ??= new Cerebras({ + apiKey: currentUser().llmConfig.cerebrasKey || process.env.CEREBRAS_API_KEY, + }); return this._client; }