-
Notifications
You must be signed in to change notification settings - Fork 35
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How can I unload a model from memory? #90
Comments
working on that right now. i managed to do it for transformers models. i will make it possible for llama.cpp models also. there will be an option to keep in memory or unload it soon. |
there is now 2 new nodes for llama.cpp LLava Optional Memory Free Simple You can use multiple of this nodes in your workflow. |
You mean, you want to use it as an llm. I can add LLM memory free version of this node also. |
Yes, like an LLM, that's right. It will be very cool to have such a node. |
there is already llm loader and llm sampler and you can use llava models as llm on them. they are currently not supporting unloading option but i will add llm versions of them also. |
LLM Optional Memory Free would be extremely useful for many tasks, I’ll be waiting, thanks for your work! 💪 |
Problem:
Currently, the model loaded into memory (VRAM) remains in it until the end of generation, even if I generate other content (pictures, for example) after generating the text.
Question:
Is it possible to unload from memory a model that is no longer used in the next nodes of the workflow?
Is there some kind of node that allows you to controllably unload a model previously loaded into memory?
If this is not possible now, can you create such a node or suggest how you can remove unnecessary models from memory using code?
Thanks for info!
The text was updated successfully, but these errors were encountered: