-
Notifications
You must be signed in to change notification settings - Fork 62
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Ollama Support #14
Comments
Not possible at the moment, but should be straightforward to implement if you'd like to give it a shot You can check |
I wonder if it's worth implementing a wrapper/abstraction layer like LiteLLM to make things more flexible?
This is what projects like
Though I'm not currently sure if/how compatible that is with the
See also:
|
@jehna Curious, what aspects of Skimming the following prompt files:
It looks like |
There's now v2 that runs on top of llama.cpp, so adding llama3 support should be even more straightforward. @0xrsydn which version of llama3 were you planning to run? I could add it in to the new version |
i think the recent one (llama3.1 8b) is great. Thanks btw |
I researched a bit about Ollama. If I'm correct, you could run Ollama locally and Humanify could connect to its API to use any model that Ollama uses. There seems to be an undocumented feature that allows passing GBNF grammars as an argument to the model: ...but judging from other open issues about the topic I'm not really sure if it works or not. But I'll give it a try! |
This seems like it's a good overarching/summarising issue; still doesn't provide full clarity yet, but links to seemingly all the related issues, and points out that now that OpenAI supports it, it's sort of become a higher priority: Based on my read of these:
It sounds like it's not currently possible to use the GBNF functionality on the current main/released version of Ollama. According to this:
It sounds like ollama currently supports JSON mode, and that is built as a GBNF grammar (presumably on top of llama.cpp's support of it), but that the ability to use a custom grammar isn't currently exposed to the end user. |
Sadly @0xdevalias is correct and what you want to do @jehna will not work unless you patch your version of Ollama with the PR that was linked above, the release version still has no support for GBNF outside of the built in json mode. Ollama still refuses to even reply to this issue for some really strange reason, I still have no idea why they won't talk about it all and simply let the PRs keep rolling in and sit there. At this point all we can do is keep pressuring them by raising issues, making noise both here and on the Discord until we can get someone to take 10 minutes and explain to us this decision to essentially completely block this feature from end users in Ollama. |
Thank you for looking into this. I just pushed |
☝️ added llama3.1 8b model support |
How do we use ollama with this sorry if this is a dumb question. |
is it possible to use llama3 via ollama rather than huggingface one?
The text was updated successfully, but these errors were encountered: