- What is OpenLocalUI
- Features
- Roadmap
- Installation
- Contributing
- License
- Support
- Contact
- Related Projects
OpenLocalUI is a Flutter-based desktop application designed for Windows and macOS users. It aims to provide a user-friendly interface for running LLMs (Large Language Models) locally without the need for complex setups like WSL or Docker containers. Taking inspiration from OpenWebUI, which offers similar functionality in a browser-based environment, OpenLocalUI brings the convenience of a native desktop app.
-
Native Desktop Experience: OpenLocalUI is designed specifically for Windows and macOS platforms, ensuring seamless integration with your operating system.
-
LLM Execution: Run OLLAMA (Open Language Learning and Modeling Architecture) based models directly from your desktop, eliminating the need for external dependencies like WSL or Docker containers.
-
MIT License: OpenLocalUI is licensed under the permissive MIT License, encouraging contributions from the community and fostering an open-source development environment.
Despite its simplicity, OpenLocalUI has enormous potential for growth and enhancement. Based on the LangChain Dart API, future updates will focus on adding more features and improving usability. Planned features include:
- β Model Customization: Enhance the ability to customize LLM models according to specific needs.
- β Image and File Embedding: Enable embedding images and files directly into the application for more versatile usage.
- βοΈ Web and Docs Search: Integrate search functionality to allow users to access and retrieve information from web pages and local documents, making it easier to pull in relevant content for interactions.
- β Model Cross Database: Maintaining our own Ollama compatible models database will allow users to browse through a wide offer of LLMs from all sources (Ollama Library, Huggingface, 3rd parties of any kind).
- β Server Configuration and Mobile App: Provide support for configuring the UI to run on different server environments, with an accompanying mobile app for easy, on-the-go access to the platform.
β οΈ OpenLocalUI-based Ecosystem: Develop an ecosystem around OpenLocalUI, including plugins, community-driven enhancements, and third-party integrations, fostering a collaborative development environment. WE NEED YOU!
OpenLocalUI requires OLLAMA to function. Just check out the latest release and download the appropriate version based on your platform; in the folder you'll find both OpenLocalUI's and OLLAMA's installers.
Contributions to OpenLocalUI are highly encouraged and welcomed. Whether you're a developer, designer, or enthusiast, there are various ways to contribute:
- Code Contributions: Help improve the application by submitting code patches, bug fixes, or new features.
- Documentation: Improve existing documentation or create new guides to help users understand and use OpenLocalUI effectively.
- Feedback and Suggestions: Share your thoughts, ideas, and feedback to help shape the future development of OpenLocalUI.
Please refer to the CONTRIBUTING.md
file for more details on how to contribute.
OpenLocalUI comes under the permissive MIT License to encourage contributions. See the LICENSE.md
file for more information.
If you will you can foster development of OpenLocalUI with a small donation. It's a symbolical action, any amount, even the smallest one, means a lot to me. Thank you for your time regardless of what you'll do!
You can join our Discord server to get help or take part in the devlopment of OpenLocalUI. If you want to acquire any additional informations you can always write to my email address.
- Tacotron.CPP: a llama.cpp inpired local accelerated inference engine to TTS and STT models for OpenLocalUI.
- gpu_info: a dart package to retieve GPU details to model suggestions to the user with more or less no reason to exist outside of OpenLocalUI.