You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Is it planned or possible to use a local LLM for processing?
I would see this variant as a possibility to significantly increase the generation speed (if the appropriate hardware is available) and also the possibility to use the model offline.
The text was updated successfully, but these errors were encountered:
There is no "plan" as such for this. However, the use of local LLMs has been in "thoughts" lately.
Regarding the speed, token generation with Mistral Nemo appears to take longer, yes. I have been contemplating to switch back to Mistral or at least provide it as an alternative.
Let me create some tasks toward this general direction.
Just added support for offline LLMs via Ollama. An environment variable needs to be set to access this mode. Detailed steps are available in the project description.
Let me know if you get a chance to try it out, and whether it works out.
Is it planned or possible to use a local LLM for processing?
I would see this variant as a possibility to significantly increase the generation speed (if the appropriate hardware is available) and also the possibility to use the model offline.
The text was updated successfully, but these errors were encountered: