-
Notifications
You must be signed in to change notification settings - Fork 2.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Which LLM models are supported? #341
Comments
i got with same question ....... where can i see the main source code .... |
+1 |
1 similar comment
+1 |
And also can we use locally deployed LLMs other than via api keys? |
same question |
+1 |
Hi! Regarding local hosting there's a very interesting conversation going on in this thread #339 |
I have tested gemma2 and llama3 with success. The only thing that does not work locally is the embeddings. There need to be a fix to accept the style of response coming from ollama when quering embeddings... Once that is fixed you will be able to run this 100% local on a personal computer... but probably need a NVidia with 24GB of VRAM like a 3090 or a Mx Mac with 32GB RAM. |
can we use local embedding? |
Can you help me with running llama 3 from the local please.. |
Here's my .env file, put it under ./ragtest dir, hope this can help you: |
this worked for me |
I test gemma2 sucess, but glm4 failed, I guess its because of the input length of the llm |
qwen2:7b fail |
Whether other LLM models are supported, such as ChatGLM and QWEN?
The text was updated successfully, but these errors were encountered: