-
Notifications
You must be signed in to change notification settings - Fork 1.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Getting 'internal server error' while running ollama_demo file for lightrag for various smaller models. NEED HELP. #498
Comments
It seems like ollama is loading your TinyLlama with Ollama's default context size (which is 2048 on Ollama). Another thing, I don't really as I had to experiment more on this but I'm wondering if the Last thing: you can set the value of Also make sure you set I hope this will already help you, but share more if you need. |
Hey @blakkd , thanks so much for your response. I am making sure that I mention the correct context size both in the script and in the model file. I haven't experimented much after making this comment as I went on vacation shortly after, but I feel I didn't encounter this error afterwards. One other frequent error that I am getting is that of LLM failing to extract any entities or relationships. I created a different issue for this but wasn't able to attach specific details so I am creating a new one. Would you be able to help with this issue? - #537 (comment) It will be so helpful for me. Thanks! |
Just to make sure: did you run For the entities extraction, I think it's likely being related. You need a pretty large context window from my experience, but it depends on the verbosity level of the model output :D Cause sometimes I had it generate for a long time when extracting! PS: I noticed my saying were confusing about the KV tip, I was meaning to set the flag to 1 in .bashrc for example |
I am trying to run the
lightrag_ollama_demo.py
file from examples folder in the GitHub repository. I have been constantly getting this error where Ollama encounters an internal server error and stops midway while doing entity extraction. I have triedLlama3.21b
,TinyLlama
,Phi
,Qwen2.5:0.5b
withnomic-embed-text
,mxbai-embed-large
andsnowflake-arctic-embed:22m
as embedding models. I have tried different combinations of LLM and the embedding models, but I get the same error for all these models. For Qwen, it did work a few times but other times I got this error again. I saw that others also got this error and some suggestions were to change theOLLAMA_KV_CACHE_TYPE
to q8_0 and others suggested that after new changes, this error has been fixed. I tried changing the KV value to q8_0 through this command -launchctl setenv OLLAMA_KV_CACHE_TYPE q8_0
in my terminal but even that didn't work. And I pulled all the recent changes only day before yesterday but I am still getting this error.Here is my Ollama log if that helps -
The text was updated successfully, but these errors were encountered: