Multi-user web cockpit for Ollama: dashboard + chat + admin, pip-installable in one process
Project description
llm-cockpit
A local-first, multi-user web interface for Ollama: a dashboard for what's loaded and how it's behaving, plus a Claude-shaped chat / code UI you can pip install and have running in five minutes.
The cockpit assumes you already have Ollama running. It does not install, manage, or supervise Ollama — it talks to it.
Status
v1.0.0 — public PyPI release. The core local cockpit is implemented,
published through PyPI trusted publishing, and smoke-tested on Neuroforge.
See docs/process/SPRINT_STATE.md and
docs/specs/functional/UC-11-pypi-publish.md.
What it does
- Dashboard with placement board. Kanban-style zones —
GPU 0,GPU 1, …,Cross GPU,On Demand. Admin drag-drops model cards to shape what's warm where; non-admin sees the board read-only. Each card is intentionally compact: 30-day calls, cold-load time, single-GPU and tensor/multi-GPU tokens/s, single-GPU and tensor/multi-GPU context, and a temperature-backed heat signal. "Load model" searches the Ollama registry and downloads without leaving the page. GPU panel is optional (nvidia-smi). - Chat. Pick any chat-tagged model from your Ollama install and have a streaming conversation. Per-user history, per-conversation system prompt, code-block highlighting.
- Code. Same shell as Chat, filtered to code-tagged models, with a coder-default system prompt and diff rendering.
- Admin (user management). Add / delete users, set roles on a
chat < code < adminladder, reset passwords. Force first-login password change for any seeded or admin-created account. - Admin (Ollama configuration). Sortable model-management table with tag, placement, keep-alive, performance metrics, per-model test/delete, and sequential "Test all models" progress/ETA. The page also contains the tagging-heuristic editor, code-mode default system prompt, per-model metrics drill-down, and full audit log.
- LAN access. Installer asks whether to bind to
127.0.0.1only or0.0.0.0, so phones / tablets / other laptops on the same LAN can use the cockpit without a reverse proxy. HTTPS is out of scope for v0.1; for off-LAN access use a VPN (Tailscale / WireGuard) or a TLS terminator.
Quick start
# 1. Have Ollama running (https://ollama.com/download)
ollama serve # or: systemctl --user start ollama
# 2. Install the cockpit from PyPI
pipx install llm-cockpit
# 3. Bootstrap (probes Ollama, creates admin / ollama, sets must_change_password)
cockpit-admin init
# 4. Run
cockpit-admin serve
# 5. Open http://localhost:8080 → log in as admin / ollama → change password → use.
Other planned shapes:
cockpit-admin systemd-installon Linux once UC-08 Slice E is re-verified.
Roles (ADR-004)
Each user has one role on a ladder. Higher roles include lower-rung capabilities.
| Role | What it can do |
|---|---|
chat |
Log in, chat with chat-tagged models, see own conversations, change own password. |
code |
Above + code with code-tagged models, see own code conversations. |
admin |
Above + manage users, configure Ollama (tags, pull/delete, defaults), see system-wide metrics + audit log. |
Bootstrap seeds one user: admin / ollama with a forced password change on first login.
Repo layout
src/cockpit/ Python package (planned shape per ADR-002 v1.1)
├── cli.py cockpit-admin entry point
├── main.py FastAPI app
├── routers/ auth, dashboard, chat, code, admin_users, admin_ollama
├── services/ users, model_tags, metrics, audit, settings
├── ports/ LLMChat, Telemetry (hexagonal)
├── adapters/ ollama_chat, telemetry, fake_chat, fake_telemetry
├── models.py / schemas.py
├── migrations/ alembic
├── frontend_dist/ built Next.js static export, bundled at wheel-build time
└── default_config/ model_tag_heuristics.yaml, code_default_system_prompt.md
docs/ mirror of the vault subset (synced at sprint review)
├── PROCESS.md, SPRINT_STATE.md
├── decisions/ ADR-001..004
├── design-principles/ DP-INDEX (inherits from AgenticBlox)
├── specs/{user,functional,test}/ UC-01..UC-12
├── architecture/COMPONENTS.md
└── STATUS.md
scripts/sync-docs-from-vault.sh
Documentation
| Where | What |
|---|---|
docs/PROCESS.md |
Spec-First + 1-week-sprint discipline. |
docs/architecture/COMPONENTS.md |
Component map + the two ports (LLMChat, Telemetry). |
docs/decisions/ |
ADRs. ADR-001 process; ADR-002 stack; ADR-003 public framing; ADR-004 role ladder. |
docs/design-principles/DP-INDEX.md |
Which AgenticBlox DPs we adopt, defer, or skip. |
docs/specs/ |
One folder per spec type (user / functional / test). |
Process
Vault is the source of truth (DP-024); docs/ is the mirror, updated at sprint review by scripts/sync-docs-from-vault.sh.
Status flow Draft → Review → Accepted → In Progress → Done → User Accepted. Implementation only starts on a Functional Spec at status Accepted. Review→Accepted and Done→User Accepted always require explicit owner approval.
Branches: feature/US-NN-short-title → develop → main. Commit prefix: [US-NN] short description.
License
MIT. See LICENSE.
Project home
This repo is the implementation. The design source-of-truth is the project hub in the Obsidian vault at 020 Projects/LLM-Cockpit/.
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file llm_cockpit-1.0.0.tar.gz.
File metadata
- Download URL: llm_cockpit-1.0.0.tar.gz
- Upload date:
- Size: 846.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
779f0329c7dcdc96d501fa3371d477ebd8f70a7bd1e8b25ac8182b2a50319b72
|
|
| MD5 |
4b9404a6a980a0a04fd4cd01b6409250
|
|
| BLAKE2b-256 |
4aa2d22e70dcf68fe18ae7183bd3adc83f9d5d963df6c141f75bb43c7e977f02
|
Provenance
The following attestation bundles were made for llm_cockpit-1.0.0.tar.gz:
Publisher:
release.yml on Bloxperts/llm-cockpit
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
llm_cockpit-1.0.0.tar.gz -
Subject digest:
779f0329c7dcdc96d501fa3371d477ebd8f70a7bd1e8b25ac8182b2a50319b72 - Sigstore transparency entry: 1429156199
- Sigstore integration time:
-
Permalink:
Bloxperts/llm-cockpit@d588007f8b9c22171fa184f5fd7737c05a9a0f14 -
Branch / Tag:
refs/tags/v1.0.0 - Owner: https://github.com/Bloxperts
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
release.yml@d588007f8b9c22171fa184f5fd7737c05a9a0f14 -
Trigger Event:
push
-
Statement type:
File details
Details for the file llm_cockpit-1.0.0-py3-none-any.whl.
File metadata
- Download URL: llm_cockpit-1.0.0-py3-none-any.whl
- Upload date:
- Size: 872.9 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
7e05cae8ea54b53ec15ce7efd82ac3efc2b5f8bc5226c869fe9895b2badeebcf
|
|
| MD5 |
2eda07713d3e3409714f95a959e626c7
|
|
| BLAKE2b-256 |
b34a657bbb5fa08e3aeba7a695c6a43ec3bf6bdf34ec890bb669caf7c5703971
|
Provenance
The following attestation bundles were made for llm_cockpit-1.0.0-py3-none-any.whl:
Publisher:
release.yml on Bloxperts/llm-cockpit
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
llm_cockpit-1.0.0-py3-none-any.whl -
Subject digest:
7e05cae8ea54b53ec15ce7efd82ac3efc2b5f8bc5226c869fe9895b2badeebcf - Sigstore transparency entry: 1429156204
- Sigstore integration time:
-
Permalink:
Bloxperts/llm-cockpit@d588007f8b9c22171fa184f5fd7737c05a9a0f14 -
Branch / Tag:
refs/tags/v1.0.0 - Owner: https://github.com/Bloxperts
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
release.yml@d588007f8b9c22171fa184f5fd7737c05a9a0f14 -
Trigger Event:
push
-
Statement type: