LLM Controller
Local GGUF chat, control, and runtime visibility. CE v1.0
LLM Controller CE app interface showing chat, model controls, runtime visibility drawers, logs, GPU monitoring, analytics, benchmarks, and administration tools
</> View on GitHub 📚 Read the Wiki 🛣️ Product Overview ⚖️ CE vs Pro 📓 Read the Blog

🧠 Model Control


Admins scan GGUF folders, govern the registry, load and stop local GGUF models, and save llama-server runtime defaults.

💬 Chat-First Workflows


Signed-in users chat with the loaded main model, stream responses, stop generation, regenerate replies, edit prompts, and manage saved chats.

📎 File-Aware Conversations


Attach supported text and code files directly to prompts, with server-side limits, chunking, context budgeting, and persistence in saved conversation turns.

🖥️ Runtime Visibility


Watch live llama-server Logs, runtime status, active process visibility, and GPU Monitor data from local NVIDIA or AMD tools where available.

📊 Analytics & Benchmarks


Review runtime Analytics and run admin-managed Benchmarks with editable prompts, live progress, best-run tracking, and result drill-downs.

⚙️ Installation & Settings


Bootstrap the Community Edition with first-run installation, runtime path settings, user administration, and local configuration.

What LLM Controller CE Includes

Get Release Updates

No spam. Just release notes, wiki updates, and important project news.