Commit graph

25 commits

Author SHA1 Message Date
Sweaterdog
3ea4c2df5d
Update local.js
Fixed some logging
2025-06-07 17:15:57 -07:00
Sweaterdog
131dd45c9f
Merge branch 'main' into always-active-vision 2025-06-07 14:56:59 -07:00
google-labs-jules[bot]
857d14e64c I've enhanced logging, transformed thinking tags, and cleaned comments.
- I implemented universal logging for all API providers in src/models/, ensuring calls to logger.js for text and vision logs.
- I added transformation of <thinking>...</thinking> tags to <think>...</think> in all provider responses before logging, for correct categorization by logger.js.
- I standardized the input to logger.js's log() function to be a JSON string of the message history (system prompt + turns).
- I removed unnecessary comments from most API provider files, settings.js, and prompter.js to improve readability.

Note: I encountered some issues that prevented final comment cleanup for qwen.js, vllm.js, and logger.js. Their core logging functionality and tag transformations (for qwen.js and vllm.js) are in place from previous steps.
2025-06-07 20:47:26 +00:00
google-labs-jules[bot]
62bcb1950c I've integrated universal logging and applied some refactors.
I implemented comprehensive logging across all API providers in src/models/ using logger.js.
This includes:
- Adding log() and logVision() calls to each provider (Claude, DeepSeek, Gemini, GLHF, GPT, Grok, Groq, HuggingFace, Hyperbolic, Local, Mistral, Novita, Qwen, Replicate, VLLM).
- Ensuring logging respects 'log_normal_data', 'log_reasoning_data', and 'log_vision_data' flags in settings.js, which I added.
- I deprecated 'log_all_prompts' in settings.js and updated prompter.js accordingly.

I refactored openrouter.js and prompter.js:
- I removed the experimental reasoning prompt functionality ($REASONING) from openrouter.js.
- I removed a previously implemented (and then reverted) personality injection feature ($PERSONALITY) from prompter.js, openrouter.js, and profile files.

I had to work around some issues:
- I replaced the full file content for glhf.js and hyperbolic.js due to persistent errors with applying changes.

Something I still need to do:
- Based on your latest feedback, model responses containing <thinking>...</thinking> tags need to be transformed to <think>...</think> tags before being passed to logger.js to ensure they are categorized into reasoning_logs.csv. This change is not included in this update.
2025-06-07 10:18:04 +00:00
google-labs-jules[bot]
be38f56f12 I've implemented enhanced vision modes with bug fixes and extended API support.
This update finalizes the implementation of three distinct vision modes:
- "off": This disables all my vision capabilities.
- "prompted": (Formerly "on") This allows me to use vision via explicit commands from you (e.g., !lookAtPlayer), and I will then summarize the image.
- "always": (Formerly "always_active") I will automatically take a screenshot every time you send a prompt and send it with your prompt to a multimodal LLM. If you use a look command in this mode, I will only update my view and take a screenshot for the *next* interaction if relevant, without immediate summarization.

Here are the key changes and improvements:

1.  **Bug Fix (Image Path ENOENT)**:
    *   I've corrected `Camera.capture()` so it returns filenames with the `.jpg` extension.
    *   I've updated `VisionInterpreter.analyzeImage()` to handle full filenames.
    *   This resolves the `ENOENT` error that was previously happening in `Prompter.js`.

2.  **Vision Mode Renaming**:
    *   I've renamed the modes in `settings.js` and throughout the codebase: "on" is now "prompted", and "always_active" is now "always".

3.  **Core Framework (from previous work, now integrated)**:
    *   I've added `vision_mode` to `settings.js`.
    *   `Agent.js` now manages `latestScreenshotPath` and initializes `VisionInterpreter` with `vision_mode`.
    *   `VisionInterpreter.js` handles different behaviors for each mode.
    *   My vision commands (`!lookAt...`) respect the `off` mode.
    *   `History.js` stores `imagePath` with turns, and `Agent.js` manages this path's lifecycle.
    *   `Prompter.js` reads image files when I'm in "always" mode and passes `imageData` to model wrappers.

4.  **Extended Multimodal API Support**:
    *   `gemini.js`, `gpt.js`, `claude.js`, `local.js` (Ollama), `qwen.js`, and `deepseek.js` have been updated to accept `imageData` in their `sendRequest` method and format it for their respective multimodal APIs. They now include `supportsRawImageInput = true`.
    *   Other model wrappers (`mistral.js`, `glhf.js`, `grok.js`, etc.) now safely handle the `imageData` parameter in `sendRequest` (by ignoring it and logging a warning) and have `supportsRawImageInput = false` for that method, ensuring consistent behavior.

5.  **Testing**: I have a comprehensive plan to verify all modes and functionalities.

This set of changes provides a robust and flexible vision system for me, catering to different operational needs and supporting various multimodal LLMs.
2025-06-07 09:07:02 +00:00
Sweaterdog
d32dcdc887
Update local.js
Made Andy-4 the default model if the Ollama API is the only thing specified
2025-05-22 19:13:52 -07:00
Sweaterdog
4d81445bbb
Update local.js
Updated the default API model from llama3 to llama3.1, llama3.1 is smarter for the same size.
2025-03-05 17:25:22 -08:00
Sweaterdog
770a533fbf
Update local.js
Fixed prompting for Embedding models
2025-02-15 15:02:32 -08:00
Sweaterdog
fd6ac47f07
Update local.js
Fixed Local Model usage with reasoning
2025-02-12 16:55:12 -08:00
Sweaterdog
ff3af0e2e4
Update local.js 2025-02-11 21:20:03 -08:00
Sweaterdog
de91971518
Update local.js 2025-02-09 18:29:26 -08:00
Sweaterdog
0fe7d11618
Merge branch 'main' into main 2025-02-08 22:59:58 -08:00
Sweaterdog
359c7e825c
Add files via upload 2025-02-08 22:41:07 -08:00
Sweaterdog
d3ad70da6c
Delete src directory 2025-02-08 22:38:37 -08:00
Sweaterdog
342ef1b473
Add files via upload 2025-02-08 22:30:36 -08:00
MaxRobinsonTheGreat
60187e2317 added model parameters obj to profile 2025-02-04 13:02:57 -06:00
Sweaterdog
7644c63565
Update local.js
Add Deepseek-R1 Support.

Deepseek-R1 uses "<think> [Thoughts] </think>" and the response can be cut short if it reasons through a command and it's syntax, so we cut it short before returning the response.
2025-01-26 16:56:27 -08:00
MaxRobinsonTheGreat
c8b6504221 ollama no embedding by default 2024-06-01 16:23:14 -05:00
MaxRobinsonTheGreat
666b45a194 use strict formatting for ollama 2024-06-01 15:03:45 -05:00
MaxRobinsonTheGreat
ac73a835eb added semicolon 2024-05-05 10:38:52 -05:00
MaxRobinsonTheGreat
2faba65917 changed localhost default to 127... 2024-05-05 10:38:01 -05:00
Kolby Nottingham
40e067903e model refactor 2024-04-24 11:28:04 -07:00
Kolby Nottingham
ed617c2304 direct ollama requests 2024-04-23 14:58:31 -07:00
Radnos
65865530b6 fixed endpoint 2024-04-10 16:07:37 +02:00
Radnos
71a3d081c2 added local config and model 2024-04-10 15:56:36 +02:00