LumiChats Offline LLM
Run powerful AI privately on your Windows PC
LumiChats Offline is a free, open-source desktop application that lets you run state-of-the-art AI language models entirely on your Windows PC — no internet connection, no cloud, no data collection. Your conversations never leave your device.
Overview
LumiChats Offline is built on top of GPT4All by Nomic AI — the leading open-source local LLM runner — with LumiChats branding, enhanced privacy defaults, a refined dark theme, and our custom fine-tuned models optimised for real-world tasks.
Privacy First
Zero data leaves your machine. No telemetry, no analytics, no opt-in prompts.
CPU Optimised
Runs efficiently on everyday hardware. No GPU, no CUDA, no Vulkan required.
LocalDocs (RAG)
Chat with your own PDFs and Word files privately using local embeddings.
LumiChats Models
Fine-tuned open-source models optimised for instruction following and coding.
Download & Install
LumiChats Offline v1.1
System Requirements
| OS | Windows 10 or Windows 11 (64-bit) |
| RAM | 8 GB minimum · 16 GB recommended |
| Storage | 5–10 GB free per model downloaded |
| GPU | Not required — CPU-only inference |
| Internet | Only needed to download models initially |
Quickstart
- 1
Download the app
Click the download button above and extract the zip folder anywhere on your PC.
- 2
Run chat.exe
Double-click chat.exe inside the extracted folder. No installation wizard needed.
- 3
Install a model
Click Models in the sidebar → + Add Model → choose and download a model (~1–5 GB).
- 4
Start chatting
Click Chats → + New Chat → select your model → type your message and press Enter.
Installing Models
LumiChats Offline supports any GGUF-format model. Models are downloaded to your device and run entirely locally. Browse and install models directly from inside the app.
| Model | Size | RAM | Best for |
|---|---|---|---|
| Qwen2 1.5B | 894 MB | 3 GB | Quick tasks, low RAM devices |
| LumiChats v1.1 | ~1.8 GB | 4 GB | General chat, instruction following |
| LumiChats Coder v2.1 | ~1.4 GB | 4 GB | Code generation & debugging |
| Phi-3 Mini | 2.18 GB | 4 GB | Balanced quality & speed |
| Mistral 7B | 4.1 GB | 8 GB | High quality general tasks |
| LLaMA 3 8B | 4.66 GB | 8 GB | Best overall quality |
LocalDocs (RAG)
LocalDocs lets you chat with your own files — PDFs, Word documents, text files — using local retrieval-augmented generation (RAG). Your files are embedded locally using the bundled nomic-embed-text model. Nothing is sent to any server.
- 1
Click LocalDocs in the sidebar
- 2
Click + Add Collection and name your collection
- 3
Select a folder on your PC that contains your documents
- 4
Click Create Collection — embedding runs locally in the background
- 5
When the green Ready indicator appears, your docs are indexed
- 6
In any chat, open LocalDocs from the top-right button to activate it
- 7
Ask questions — the model will reference your documents for answers
Settings Reference
Application Settings
| Setting | Description | Default |
|---|---|---|
| Theme | Light, Dark, or LegacyDark. LumiChats defaults to Dark. | Dark |
| Font Size | Small, Medium, or Large | Small |
| CPU Threads | Number of CPU threads for inference. More = faster on multi-core CPUs. | Auto |
| Enable Local Server | Expose an OpenAI-compatible API on localhost:4891 | Off |
Model Settings
| Setting | Description | Default |
|---|---|---|
| Context Length | Max token window size | 2048 |
| Temperature | Higher = more creative. Lower = more factual. | 0.7 |
| Max Response Length | Maximum tokens in a single response | 4096 |
LocalDocs Settings
| Setting | Description | Default |
|---|---|---|
| Document Snippet Size | Characters per indexed snippet | 512 |
| Max Snippets Per Prompt | How many file snippets the model sees per turn | 3 |
| Show Sources | Display which files were referenced below each response | On |
Build from Source
For developers who want to build LumiChats Offline from the source code. Requires Windows with Visual Studio 2022.
Prerequisites
- Visual Studio 2022 with Desktop C++ workload
- CMake 3.21+
- Qt 6.8+ (installed via aqtinstall)
- Python 3.10+
Clone & Build
git clone https://github.com/adityajhakumar/LumiChats-Offline-LLM.git cd LumiChats-Offline-LLM # Install Qt pip install aqtinstall aqt install-qt windows desktop 6.8.3 win64_msvc2022_64 -O C:\Qt # Configure (run in x64 Native Tools Command Prompt) cmake -S gpt4all-chat -B build ^ -DCMAKE_BUILD_TYPE=Release ^ -DCMAKE_PREFIX_PATH="C:\Qt\6.8.3\msvc2022_64" ^ -DLLMODEL_CUDA=OFF -DLLMODEL_VULKAN=OFF cmake --build build --config Release --parallel # Deploy Qt dependencies C:\Qt\6.8.3\msvc2022_64\bin\windeployqt.exe build\bin\Release\chat.exe
LumiChats Models
We have fine-tuned a suite of open-source models specifically for LumiChats use cases. All models are free to use on Hugging Face and can be downloaded into LumiChats Offline via the Models tab.
FAQ
Does LumiChats Offline send any data to the internet?
No. Once models are downloaded, the app runs entirely offline. Your conversations, documents, and data never leave your device. All telemetry and data-sharing features are disabled by default.
Do I need a GPU?
No. LumiChats Offline is optimised for CPU-only inference. It runs well on everyday Windows laptops and desktops without any dedicated GPU.
What model formats are supported?
LumiChats Offline supports GGUF models — the standard format used by llama.cpp. You can download models directly inside the app or sideload any GGUF model file.
Can I use it alongside LumiChats web?
Yes. LumiChats Offline and LumiChats web are independent products. Use Offline for maximum privacy or when you have no internet, and web for access to frontier models like Claude, GPT-5, and Gemini.
Is this based on GPT4All?
Yes. LumiChats Offline is built on top of GPT4All v3.10 by Nomic AI, used under the MIT License. We have applied LumiChats branding, privacy enhancements, and integrated our own fine-tuned models.
Attribution & License
LumiChats Offline is a derivative work of GPT4All by Nomic AI and contributors, used under the MIT License. We are deeply grateful to the Nomic AI team for building and open-sourcing this foundational technology.
@misc{gpt4all,
author = {Anand, Yuvanesh and Nussbaum, Zach and Duderstadt, Brandon
and Schmidt, Benno and others},
title = {GPT4All: An Ecosystem of Open Source Compressed LLMs},
year = {2023},
publisher = {Nomic AI},
url = {https://github.com/nomic-ai/gpt4all}
}LumiChats-specific modifications © 2026 LumiChats · Original GPT4All code retains its copyright by Nomic AI and contributors. Model weights downloaded through the app are subject to their respective licenses.
Ready to get started?
Download LumiChats Offline and run AI privately on your Windows PC — free, forever.
↓ Download for Windows — Free