-
Notifications
You must be signed in to change notification settings - Fork 921
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Some questions regarding the WebLLM tool #639
Comments
(just saw your issue, just a web-llm user too ) local model ala ollama would be significantly faster than their web counter part.
Done this way it seems it would be an optimisation. i would love it too |
(I'm not a maintainer)
Sorta. It runs LLMs in the browser (not native). Charlie Ruan descibes the architecture in more detail in this video: https://www.youtube.com/watch?v=MhTCzq7iTy0&pp=ygUMd2ViYWkgd2VibGxt
That depends on your definition of "local". The models are retrieved from a server on the Internet (not local). They are then cached to browser persistent storage and can be executed entirely offline (local).
By design, user prompts and responses aren't sent anywhere outside your device. It's good to be vigilant as a developer though. For example a supply chain attack in one of WebLLM's dependencies might add a backdoor that sends user prompts to a capturing endpoint. The community would probably figure that out pretty quickly.
No, and that seems outside of its mission to provide local LLM functionality. |
Yeah, that's what I did with my project (create-decent-app). I think it's a good approach though it's a little difficult to know you have the exact same model in Ollama or WebLLM. |
Thank you for your response! However, I still have a few questions, such as: If WebLLM uses default models from the internet, can I still use it completely offline after disconnecting from the network? Can WebLLM utilize locally downloaded GGUF models from HuggingFace, similar to tools like Ollama or LM Studio? In what scenarios would using WebLLM be the preferred choice? Otherwise, wouldn't it be better to directly use Ollama for running LLMs locally, especially when combined with OpenWebUI? Thank you! |
Dear WebLLM Team,
I recently came across a Facebook group where someone shared your achievements, and I found it very intriguing. However, I have a few questions about WebLLM that I’d like to ask:
Which country or company/team developed WebLLM?
Does WebLLM work based on the principle of running LLMs locally on a computer? However, it seems to allow the use of WebLLM's default models. In this case, is it considered local operation? What about the privacy risks of conversation data?
Can WebLLM utilize locally downloaded GGUF models, similar to tools like Ollama or LM Studio?
Additionally, does it support integration with OpenAI API keys for usage?
Thank you!
The text was updated successfully, but these errors were encountered: