diff --git a/README.md b/README.md index b1ff8cb..191b21a 100644 --- a/README.md +++ b/README.md @@ -65,7 +65,7 @@ You can configure the agent's name, model, and prompts in their profile like `an | `hyperbolic` | `HYPERBOLIC_API_KEY` | `hyperbolic/deepseek-ai/DeepSeek-V3` | [docs](https://docs.hyperbolic.xyz/docs/getting-started) | | `vllm` | n/a | `vllm/llama3` | n/a | | `cerebras` | `CEREBRAS_API_KEY` | `cerebras/llama-3.3-70b` | [docs](https://inference-docs.cerebras.ai/introduction) | -| `mercury(EA)` | `MERCURY_API_KEY` | `mercury-coder-small` | [docs](https://www.inceptionlabs.ai/) | +| `mercury` | `MERCURY_API_KEY` | `mercury-coder-small` | [docs](https://www.inceptionlabs.ai/) | If you use Ollama, to install the models used by default (generation and embedding), execute the following terminal command: `ollama pull llama3.1 && ollama pull nomic-embed-text` diff --git a/profiles/mercury.json b/profiles/mercury.json index cac6d49..482b601 100644 --- a/profiles/mercury.json +++ b/profiles/mercury.json @@ -3,13 +3,7 @@ "cooldown": 5000, - "model": { - "api": "mercury", - "url": "https://api.inceptionlabs.ai/v1", - "model": "mercury-coder-small" - }, + "model": "mercury/mercury-coder-small", - "embedding": "openai", - - "description":"Official Website Introduction:The world’s first diffusion large language models" + "embedding": "openai" } \ No newline at end of file diff --git a/src/models/mercury.js b/src/models/mercury.js index d4c4c3e..74cd64e 100644 --- a/src/models/mercury.js +++ b/src/models/mercury.js @@ -3,12 +3,15 @@ import { getKey, hasKey } from '../utils/keys.js'; import { strictFormat } from '../utils/text.js'; export class Mercury { + static prefix = 'mercury'; constructor(model_name, url, params) { this.model_name = model_name; this.params = params; let config = {}; if (url) config.baseURL = url; + else + config.baseURL = "https://api.inceptionlabs.ai/v1"; config.apiKey = getKey('MERCURY_API_KEY');