A simple-to-use, open source GUI for local AI chat.
Supported platforms: iOS, Android (un-accelerated), Windows, MacOS, Linux
- Based on llama.cpp and supports any model supported by that project in GGUF form.
- Graphical user interface, tailored towards chatting with AI, which is mobile compatible.
- Create multiple chat logs, each with their own parameters, story context and characters
- Customizable character portrait images.
- Automatic model downloading from HuggingFace.
- Built-in 'narrator' using the slash-command '/narrator' at the start of a message.
- Lorebook support: customizable lists of 'lore' text that gets added behind the scenes to the LLM prompt when a pattern is matched in the chat log context or recent messages. Lorebooks are active across all chats and are enabled if a character name from the chat log pattern matches.
- Fast regeneration of AI chat replies with prompt caching.
- Streaming responses so you can see the AI generate it's response in real-time.
- Extensive set of AI samplers: temp, top-k, top-p, min-p, tail free sampling, locally typical sampling, dynamic temperature, DRY and XTC.
See the documentation for the quick start guide.
See the documentation for the build instructions for all supported platforms.
Currently shipping the first official release binaries off to the right hand side of the main page or via that Link.
The author is new at shipping apps with more complicated, non-static library dependencies, the first attempt at deploying an AppImage for Linux builds and this is the first major deployment of a Flutter app for them, so if you try a release binary and it doesn't work, please create an issue or start a discussion!
This project is licensed under the GPL v3 terms, as specified in the LICENSE
file.
The project is built around woolycore, woolydart for the language bindings and of course the great llama.cpp library. All three of these libraries are licensed under the MIT license.