LumiChats Offline LLM

Run powerful AI privately on your Windows PC

LumiChats Offline is a free, open-source desktop application that lets you run state-of-the-art AI language models entirely on your Windows PC — no internet connection, no cloud, no data collection. Your conversations never leave your device.

Free & Open SourceNo GPU Required100% PrivateWindows 10 / 11No Internet Needed

Overview

LumiChats Offline is built on top of GPT4All by Nomic AI — the leading open-source local LLM runner — with LumiChats branding, enhanced privacy defaults, a refined dark theme, and our custom fine-tuned models optimised for real-world tasks.

Privacy First

Zero data leaves your machine. No telemetry, no analytics, no opt-in prompts.

CPU Optimised

Runs efficiently on everyday hardware. No GPU, no CUDA, no Vulkan required.

LocalDocs (RAG)

Chat with your own PDFs and Word files privately using local embeddings.

LumiChats Models

Fine-tuned open-source models optimised for instruction following and coding.

Download & Install

LumiChats Offline v1.1

Windows 10 / 11 · 64-bit · ~200 MB

↓ Download for Windows

System Requirements

OSWindows 10 or Windows 11 (64-bit)
RAM8 GB minimum · 16 GB recommended
Storage5–10 GB free per model downloaded
GPUNot required — CPU-only inference
InternetOnly needed to download models initially

Quickstart

  1. 1

    Download the app

    Click the download button above and extract the zip folder anywhere on your PC.

  2. 2

    Run chat.exe

    Double-click chat.exe inside the extracted folder. No installation wizard needed.

  3. 3

    Install a model

    Click Models in the sidebar → + Add Model → choose and download a model (~1–5 GB).

  4. 4

    Start chatting

    Click Chats → + New Chat → select your model → type your message and press Enter.

Installing Models

LumiChats Offline supports any GGUF-format model. Models are downloaded to your device and run entirely locally. Browse and install models directly from inside the app.

ModelSizeRAMBest for
Qwen2 1.5B894 MB3 GBQuick tasks, low RAM devices
LumiChats v1.1~1.8 GB4 GBGeneral chat, instruction following
LumiChats Coder v2.1~1.4 GB4 GBCode generation & debugging
Phi-3 Mini2.18 GB4 GBBalanced quality & speed
Mistral 7B4.1 GB8 GBHigh quality general tasks
LLaMA 3 8B4.66 GB8 GBBest overall quality

LocalDocs (RAG)

LocalDocs lets you chat with your own files — PDFs, Word documents, text files — using local retrieval-augmented generation (RAG). Your files are embedded locally using the bundled nomic-embed-text model. Nothing is sent to any server.

  1. 1

    Click LocalDocs in the sidebar

  2. 2

    Click + Add Collection and name your collection

  3. 3

    Select a folder on your PC that contains your documents

  4. 4

    Click Create Collection — embedding runs locally in the background

  5. 5

    When the green Ready indicator appears, your docs are indexed

  6. 6

    In any chat, open LocalDocs from the top-right button to activate it

  7. 7

    Ask questions — the model will reference your documents for answers

Tip: Supported file types include .txt, .pdf, .md, .rst, and .docx. For best results, use phrases like "in the docs" or "from the provided files" when prompting.

Settings Reference

Application Settings

SettingDescriptionDefault
ThemeLight, Dark, or LegacyDark. LumiChats defaults to Dark.Dark
Font SizeSmall, Medium, or LargeSmall
CPU ThreadsNumber of CPU threads for inference. More = faster on multi-core CPUs.Auto
Enable Local ServerExpose an OpenAI-compatible API on localhost:4891Off

Model Settings

SettingDescriptionDefault
Context LengthMax token window size2048
TemperatureHigher = more creative. Lower = more factual.0.7
Max Response LengthMaximum tokens in a single response4096

LocalDocs Settings

SettingDescriptionDefault
Document Snippet SizeCharacters per indexed snippet512
Max Snippets Per PromptHow many file snippets the model sees per turn3
Show SourcesDisplay which files were referenced below each responseOn

Build from Source

For developers who want to build LumiChats Offline from the source code. Requires Windows with Visual Studio 2022.

Prerequisites

  • Visual Studio 2022 with Desktop C++ workload
  • CMake 3.21+
  • Qt 6.8+ (installed via aqtinstall)
  • Python 3.10+

Clone & Build

git clone https://github.com/adityajhakumar/LumiChats-Offline-LLM.git
cd LumiChats-Offline-LLM

# Install Qt
pip install aqtinstall
aqt install-qt windows desktop 6.8.3 win64_msvc2022_64 -O C:\Qt

# Configure (run in x64 Native Tools Command Prompt)
cmake -S gpt4all-chat -B build ^
  -DCMAKE_BUILD_TYPE=Release ^
  -DCMAKE_PREFIX_PATH="C:\Qt\6.8.3\msvc2022_64" ^
  -DLLMODEL_CUDA=OFF -DLLMODEL_VULKAN=OFF

cmake --build build --config Release --parallel

# Deploy Qt dependencies
C:\Qt\6.8.3\msvc2022_64\bin\windeployqt.exe build\bin\Release\chat.exe

LumiChats Models

We have fine-tuned a suite of open-source models specifically for LumiChats use cases. All models are free to use on Hugging Face and can be downloaded into LumiChats Offline via the Models tab.

FAQ

Does LumiChats Offline send any data to the internet?

No. Once models are downloaded, the app runs entirely offline. Your conversations, documents, and data never leave your device. All telemetry and data-sharing features are disabled by default.

Do I need a GPU?

No. LumiChats Offline is optimised for CPU-only inference. It runs well on everyday Windows laptops and desktops without any dedicated GPU.

What model formats are supported?

LumiChats Offline supports GGUF models — the standard format used by llama.cpp. You can download models directly inside the app or sideload any GGUF model file.

Can I use it alongside LumiChats web?

Yes. LumiChats Offline and LumiChats web are independent products. Use Offline for maximum privacy or when you have no internet, and web for access to frontier models like Claude, GPT-5, and Gemini.

Is this based on GPT4All?

Yes. LumiChats Offline is built on top of GPT4All v3.10 by Nomic AI, used under the MIT License. We have applied LumiChats branding, privacy enhancements, and integrated our own fine-tuned models.

Attribution & License

LumiChats Offline is a derivative work of GPT4All by Nomic AI and contributors, used under the MIT License. We are deeply grateful to the Nomic AI team for building and open-sourcing this foundational technology.

@misc{gpt4all,
  author    = {Anand, Yuvanesh and Nussbaum, Zach and Duderstadt, Brandon
               and Schmidt, Benno and others},
  title     = {GPT4All: An Ecosystem of Open Source Compressed LLMs},
  year      = {2023},
  publisher = {Nomic AI},
  url       = {https://github.com/nomic-ai/gpt4all}
}

LumiChats-specific modifications © 2026 LumiChats · Original GPT4All code retains its copyright by Nomic AI and contributors. Model weights downloaded through the app are subject to their respective licenses.

Ready to get started?

Download LumiChats Offline and run AI privately on your Windows PC — free, forever.

↓ Download for Windows — Free