Local AI Dashboard
Operate local GGUF models with llama-server control.
Live runtime visibility, no cloud dependency.
CE v1.0

๐Ÿง  Model Management


Admins scan GGUF folders, govern the registry, start and stop model processes, and save runtime defaults for the workspace.

๐Ÿ’ฌ Modern Chat Workflows


Signed-in users chat with the loaded main model, stream responses, stop generation, regenerate replies, edit prompts, and manage saved chats.

๐Ÿ“Ž File-Aware Conversations


Attach supported text and code files directly to prompts, with server-side limits, chunking, context budgeting, and persistence in saved conversation turns.

๐Ÿ–ฅ๏ธ Large Model Ready


Runs through a configured llama.cpp / llama-server build, supports single-file and split GGUF sets, CPU-only loading, GPU offload, tensor-split multi-GPU launch settings, Windows, and Linux.

๐Ÿ“Š Built-In Observability


Watch live logs, runtime status, analytics, NVIDIA telemetry through nvidia-smi, and AMD visibility through local ROCm tools where available.

๐Ÿงช Benchmarking Included


Run admin-managed CE benchmarks against eligible local models, edit the five-question prompt set, track live progress, and inspect best runs per model.
๐Ÿ›ฃ๏ธ Product Overview โš–๏ธ CE vs Pro ๐Ÿ““ Read the Blog

What LLM Controller CE Includes

Get Notified at Launch

No spam. Just release updates, platform milestones, and the important stuff.