diff --git a/src/lib/providers/groq.ts b/src/lib/providers/groq.ts index 62ca2d75..41004ecc 100644 --- a/src/lib/providers/groq.ts +++ b/src/lib/providers/groq.ts @@ -9,12 +9,12 @@ export const loadGroqChatModels = async () => { try { const chatModels = { - 'llama-3.2-3b-preview': { - displayName: 'Llama 3.2 3B', + 'llama-3.3-70b-versatile': { + displayName: 'Llama 3.3 70B', model: new ChatOpenAI( { openAIApiKey: groqApiKey, - modelName: 'llama-3.2-3b-preview', + modelName: 'llama-3.3-70b-versatile', temperature: 0.7, }, { @@ -22,12 +22,12 @@ export const loadGroqChatModels = async () => { }, ), }, - 'llama-3.2-11b-vision-preview': { - displayName: 'Llama 3.2 11B Vision', + 'llama-3.2-3b-preview': { + displayName: 'Llama 3.2 3B', model: new ChatOpenAI( { openAIApiKey: groqApiKey, - modelName: 'llama-3.2-11b-vision-preview', + modelName: 'llama-3.2-3b-preview', temperature: 0.7, }, { @@ -35,12 +35,12 @@ export const loadGroqChatModels = async () => { }, ), }, - 'llama-3.2-90b-vision-preview': { - displayName: 'Llama 3.2 90B Vision', + 'llama-3.2-11b-vision-preview': { + displayName: 'Llama 3.2 11B Vision', model: new ChatOpenAI( { openAIApiKey: groqApiKey, - modelName: 'llama-3.2-90b-vision-preview', + modelName: 'llama-3.2-11b-vision-preview', temperature: 0.7, }, { @@ -48,12 +48,12 @@ export const loadGroqChatModels = async () => { }, ), }, - 'llama-3.1-70b-versatile': { - displayName: 'Llama 3.1 70B', + 'llama-3.2-90b-vision-preview': { + displayName: 'Llama 3.2 90B Vision', model: new ChatOpenAI( { openAIApiKey: groqApiKey, - modelName: 'llama-3.1-70b-versatile', + modelName: 'llama-3.2-90b-vision-preview', temperature: 0.7, }, { @@ -113,19 +113,6 @@ export const loadGroqChatModels = async () => { }, ), }, - 'gemma-7b-it': { - displayName: 'Gemma 7B', - model: new ChatOpenAI( - { - openAIApiKey: groqApiKey, - modelName: 'gemma-7b-it', - temperature: 0.7, - }, - { - baseURL: 'https://api.groq.com/openai/v1', - }, - ), - }, 'gemma2-9b-it': { displayName: 'Gemma2 9B', model: new ChatOpenAI(