Skip to content

OpenLocalUI: Native desktop app for Windows, MacOS and Linux. Easily run Large Language Models locally, no complex setups required. Inspired by OpenWebUI's simplicity for LLM use.

License

Notifications You must be signed in to change notification settings

WilliamKarolDiCioccio/open_local_ui

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

OpenLocalUI

coverage build

See more screenshots

Table of Contents

  1. What is OpenLocalUI
  2. Features
  3. Roadmap
  4. Installation
  5. Contributing
  6. License
  7. Support
  8. Contact
  9. Related Projects

πŸš€ What is OpenLocalUI

OpenLocalUI is a Flutter-based desktop application designed for Windows and macOS users. It aims to provide a user-friendly interface for running LLMs (Large Language Models) locally without the need for complex setups like WSL or Docker containers. Taking inspiration from OpenWebUI, which offers similar functionality in a browser-based environment, OpenLocalUI brings the convenience of a native desktop app.

πŸ”₯ Features

  1. Native Desktop Experience: OpenLocalUI is designed specifically for Windows and macOS platforms, ensuring seamless integration with your operating system.

  2. LLM Execution: Run OLLAMA (Open Language Learning and Modeling Architecture) based models directly from your desktop, eliminating the need for external dependencies like WSL or Docker containers.

  3. MIT License: OpenLocalUI is licensed under the permissive MIT License, encouraging contributions from the community and fostering an open-source development environment.

πŸ›£οΈ Roadmap

Despite its simplicity, OpenLocalUI has enormous potential for growth and enhancement. Based on the LangChain Dart API, future updates will focus on adding more features and improving usability. Planned features include:

  1. βœ… Model Customization: Enhance the ability to customize LLM models according to specific needs.
  2. βœ… Image and File Embedding: Enable embedding images and files directly into the application for more versatile usage.
  3. βš’οΈ Web and Docs Search: Integrate search functionality to allow users to access and retrieve information from web pages and local documents, making it easier to pull in relevant content for interactions.
  4. βœ… Model Cross Database: Maintaining our own Ollama compatible models database will allow users to browse through a wide offer of LLMs from all sources (Ollama Library, Huggingface, 3rd parties of any kind).
  5. ❌ Server Configuration and Mobile App: Provide support for configuring the UI to run on different server environments, with an accompanying mobile app for easy, on-the-go access to the platform.
  6. ⚠️ OpenLocalUI-based Ecosystem: Develop an ecosystem around OpenLocalUI, including plugins, community-driven enhancements, and third-party integrations, fostering a collaborative development environment. WE NEED YOU!

πŸ’» Installation

OpenLocalUI requires OLLAMA to function. Just check out the latest release and download the appropriate version based on your platform; in the folder you'll find both OpenLocalUI's and OLLAMA's installers.

πŸ“ Contributing

Contributions to OpenLocalUI are highly encouraged and welcomed. Whether you're a developer, designer, or enthusiast, there are various ways to contribute:

  • Code Contributions: Help improve the application by submitting code patches, bug fixes, or new features.
  • Documentation: Improve existing documentation or create new guides to help users understand and use OpenLocalUI effectively.
  • Feedback and Suggestions: Share your thoughts, ideas, and feedback to help shape the future development of OpenLocalUI.

Please refer to the CONTRIBUTING.md file for more details on how to contribute.

πŸ“ƒ License

OpenLocalUI comes under the permissive MIT License to encourage contributions. See the LICENSE.md file for more information.

πŸ’– Support

Buy Me A Coffee

If you will you can foster development of OpenLocalUI with a small donation. It's a symbolical action, any amount, even the smallest one, means a lot to me. Thank you for your time regardless of what you'll do!

πŸ—¨οΈ Contact

You can join our Discord server to get help or take part in the devlopment of OpenLocalUI. If you want to acquire any additional informations you can always write to my email address.

βš’οΈ Related Projects

  • Tacotron.CPP: a llama.cpp inpired local accelerated inference engine to TTS and STT models for OpenLocalUI.
  • gpu_info: a dart package to retieve GPU details to model suggestions to the user with more or less no reason to exist outside of OpenLocalUI.

About

OpenLocalUI: Native desktop app for Windows, MacOS and Linux. Easily run Large Language Models locally, no complex setups required. Inspired by OpenWebUI's simplicity for LLM use.

Topics

Resources

License

Code of conduct

Stars

Watchers

Forks

Sponsor this project