From 6c8382874da4a628dc5d6271d3e3f522eef8c851 Mon Sep 17 00:00:00 2001 From: MaxRobinsonTheGreat Date: Sat, 23 Aug 2025 14:33:26 -0500 Subject: [PATCH] fix grok, updated many default models --- profiles/claude.json | 2 +- profiles/claude_thinker.json | 2 +- profiles/gemini.json | 2 +- profiles/grok.json | 2 +- src/models/gemini.js | 2 +- src/models/grok.js | 9 ++++----- src/models/groq.js | 2 +- src/models/novita.js | 2 +- 8 files changed, 11 insertions(+), 12 deletions(-) diff --git a/profiles/claude.json b/profiles/claude.json index d4ce4cc..27f610d 100644 --- a/profiles/claude.json +++ b/profiles/claude.json @@ -1,7 +1,7 @@ { "name": "claude", - "model": "claude-3-5-sonnet-latest", + "model": "claude-4-sonnet-latest", "embedding": "openai" } \ No newline at end of file diff --git a/profiles/claude_thinker.json b/profiles/claude_thinker.json index fdddb59..64e3fb4 100644 --- a/profiles/claude_thinker.json +++ b/profiles/claude_thinker.json @@ -2,7 +2,7 @@ "name": "claude_thinker", "model": { - "model": "claude-3-7-sonnet-latest", + "model": "claude-4-sonnet-latest", "params": { "thinking": { "type": "enabled", diff --git a/profiles/gemini.json b/profiles/gemini.json index 8a91387..0e388c7 100644 --- a/profiles/gemini.json +++ b/profiles/gemini.json @@ -1,7 +1,7 @@ { "name": "gemini", - "model": "gemini-2.0-flash", + "model": "gemini-2.5-flash", "cooldown": 5000 } diff --git a/profiles/grok.json b/profiles/grok.json index eeb3a38..eda1aaa 100644 --- a/profiles/grok.json +++ b/profiles/grok.json @@ -1,7 +1,7 @@ { "name": "Grok", - "model": "grok-beta", + "model": "grok-3-mini-latest", "embedding": "openai" } \ No newline at end of file diff --git a/src/models/gemini.js b/src/models/gemini.js index 75a20e0..ba24072 100644 --- a/src/models/gemini.js +++ b/src/models/gemini.js @@ -37,7 +37,7 @@ export class Gemini { async sendRequest(turns, systemMessage) { let model; const modelConfig = { - model: this.model_name || "gemini-1.5-flash", + model: this.model_name || "gemini-2.5-flash", // systemInstruction does not work bc google is trash }; if (this.url) { diff --git a/src/models/grok.js b/src/models/grok.js index 0753f10..40c63ce 100644 --- a/src/models/grok.js +++ b/src/models/grok.js @@ -3,7 +3,7 @@ import { getKey } from '../utils/keys.js'; // xAI doesn't supply a SDK for their models, but fully supports OpenAI and Anthropic SDKs export class Grok { - static prefix = 'grok'; + static prefix = 'xai'; constructor(model_name, url, params) { this.model_name = model_name; this.url = url; @@ -20,13 +20,12 @@ export class Grok { this.openai = new OpenAIApi(config); } - async sendRequest(turns, systemMessage, stop_seq='***') { + async sendRequest(turns, systemMessage) { let messages = [{'role': 'system', 'content': systemMessage}].concat(turns); const pack = { - model: this.model_name || "grok-beta", + model: this.model_name || "grok-3-mini-latest", messages, - stop: [stop_seq], ...(this.params || {}) }; @@ -43,7 +42,7 @@ export class Grok { catch (err) { if ((err.message == 'Context length exceeded' || err.code == 'context_length_exceeded') && turns.length > 1) { console.log('Context length exceeded, trying again with shorter context.'); - return await this.sendRequest(turns.slice(1), systemMessage, stop_seq); + return await this.sendRequest(turns.slice(1), systemMessage); } else if (err.message.includes('The model expects a single `text` element per message.')) { console.log(err); res = 'Vision is only supported by certain models.'; diff --git a/src/models/groq.js b/src/models/groq.js index 9da88c7..85a913e 100644 --- a/src/models/groq.js +++ b/src/models/groq.js @@ -50,7 +50,7 @@ export class GroqCloudAPI { let completion = await this.groq.chat.completions.create({ "messages": messages, - "model": this.model_name || "llama-3.3-70b-versatile", + "model": this.model_name || "qwen/qwen3-32b", "stream": false, "stop": stop_seq, ...(this.params || {}) diff --git a/src/models/novita.js b/src/models/novita.js index 46140f0..18e1fc4 100644 --- a/src/models/novita.js +++ b/src/models/novita.js @@ -26,7 +26,7 @@ export class Novita { messages = strictFormat(messages); const pack = { - model: this.model_name || "meta-llama/llama-3.1-70b-instruct", + model: this.model_name || "meta-llama/llama-4-scout-17b-16e-instruct", messages, stop: [stop_seq], ...(this.params || {})