Skip to main content

Multi-user web cockpit for Ollama: dashboard + chat + admin, pip-installable in one process

Project description

llm-cockpit

A local-first, multi-user web interface for Ollama: a dashboard for what's loaded and how it's behaving, plus a Claude-shaped chat / code UI you can pip install and have running in five minutes.

The cockpit assumes you already have Ollama running. It does not install, manage, or supervise Ollama — it talks to it.

Status

v1.0.0 — public PyPI release. The core local cockpit is implemented, published through PyPI trusted publishing, and smoke-tested on Neuroforge. See docs/process/SPRINT_STATE.md and docs/specs/functional/UC-11-pypi-publish.md.

What it does

  • Dashboard with placement board. Kanban-style zones — GPU 0, GPU 1, …, Cross GPU, On Demand. Admin drag-drops model cards to shape what's warm where; non-admin sees the board read-only. Each card is intentionally compact: 30-day calls, cold-load time, single-GPU and tensor/multi-GPU tokens/s, single-GPU and tensor/multi-GPU context, and a temperature-backed heat signal. "Load model" searches the Ollama registry and downloads without leaving the page. GPU panel is optional (nvidia-smi).
  • Chat. Pick any chat-tagged model from your Ollama install and have a streaming conversation. Per-user history, per-conversation system prompt, code-block highlighting.
  • Code. Same shell as Chat, filtered to code-tagged models, with a coder-default system prompt and diff rendering.
  • Admin (user management). Add / delete users, set roles on a chat < code < admin ladder, reset passwords. Force first-login password change for any seeded or admin-created account.
  • Admin (Ollama configuration). Sortable model-management table with tag, placement, keep-alive, performance metrics, per-model test/delete, and sequential "Test all models" progress/ETA. The page also contains the tagging-heuristic editor, code-mode default system prompt, per-model metrics drill-down, and full audit log.
  • LAN access. Installer asks whether to bind to 127.0.0.1 only or 0.0.0.0, so phones / tablets / other laptops on the same LAN can use the cockpit without a reverse proxy. HTTPS is out of scope for v0.1; for off-LAN access use a VPN (Tailscale / WireGuard) or a TLS terminator.

Quick start

# 1. Have Ollama running (https://ollama.com/download)
ollama serve   # or: systemctl --user start ollama

# 2. Install the cockpit from PyPI
pipx install llm-cockpit

# 3. Bootstrap (probes Ollama, creates admin / ollama, sets must_change_password)
cockpit-admin init

# 4. Run
cockpit-admin serve

# 5. Open http://localhost:8080  → log in as admin / ollama → change password → use.

Other planned shapes:

  • cockpit-admin systemd-install on Linux once UC-08 Slice E is re-verified.

Roles (ADR-004)

Each user has one role on a ladder. Higher roles include lower-rung capabilities.

Role What it can do
chat Log in, chat with chat-tagged models, see own conversations, change own password.
code Above + code with code-tagged models, see own code conversations.
admin Above + manage users, configure Ollama (tags, pull/delete, defaults), see system-wide metrics + audit log.

Bootstrap seeds one user: admin / ollama with a forced password change on first login.

Repo layout

src/cockpit/                 Python package (planned shape per ADR-002 v1.1)
├── cli.py                   cockpit-admin entry point
├── main.py                  FastAPI app
├── routers/                 auth, dashboard, chat, code, admin_users, admin_ollama
├── services/                users, model_tags, metrics, audit, settings
├── ports/                   LLMChat, Telemetry        (hexagonal)
├── adapters/                ollama_chat, telemetry, fake_chat, fake_telemetry
├── models.py / schemas.py
├── migrations/              alembic
├── frontend_dist/           built Next.js static export, bundled at wheel-build time
└── default_config/          model_tag_heuristics.yaml, code_default_system_prompt.md
docs/                        mirror of the vault subset (synced at sprint review)
├── PROCESS.md, SPRINT_STATE.md
├── decisions/               ADR-001..004
├── design-principles/       DP-INDEX (inherits from AgenticBlox)
├── specs/{user,functional,test}/  UC-01..UC-12
├── architecture/COMPONENTS.md
└── STATUS.md
scripts/sync-docs-from-vault.sh

Documentation

Where What
docs/PROCESS.md Spec-First + 1-week-sprint discipline.
docs/architecture/COMPONENTS.md Component map + the two ports (LLMChat, Telemetry).
docs/decisions/ ADRs. ADR-001 process; ADR-002 stack; ADR-003 public framing; ADR-004 role ladder.
docs/design-principles/DP-INDEX.md Which AgenticBlox DPs we adopt, defer, or skip.
docs/specs/ One folder per spec type (user / functional / test).

Process

Vault is the source of truth (DP-024); docs/ is the mirror, updated at sprint review by scripts/sync-docs-from-vault.sh.

Status flow Draft → Review → Accepted → In Progress → Done → User Accepted. Implementation only starts on a Functional Spec at status Accepted. Review→Accepted and Done→User Accepted always require explicit owner approval.

Branches: feature/US-NN-short-titledevelopmain. Commit prefix: [US-NN] short description.

License

MIT. See LICENSE.

Project home

This repo is the implementation. The design source-of-truth is the project hub in the Obsidian vault at 020 Projects/LLM-Cockpit/.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

llm_cockpit-1.0.0.tar.gz (846.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

llm_cockpit-1.0.0-py3-none-any.whl (872.9 kB view details)

Uploaded Python 3

File details

Details for the file llm_cockpit-1.0.0.tar.gz.

File metadata

  • Download URL: llm_cockpit-1.0.0.tar.gz
  • Upload date:
  • Size: 846.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for llm_cockpit-1.0.0.tar.gz
Algorithm Hash digest
SHA256 779f0329c7dcdc96d501fa3371d477ebd8f70a7bd1e8b25ac8182b2a50319b72
MD5 4b9404a6a980a0a04fd4cd01b6409250
BLAKE2b-256 4aa2d22e70dcf68fe18ae7183bd3adc83f9d5d963df6c141f75bb43c7e977f02

See more details on using hashes here.

Provenance

The following attestation bundles were made for llm_cockpit-1.0.0.tar.gz:

Publisher: release.yml on Bloxperts/llm-cockpit

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file llm_cockpit-1.0.0-py3-none-any.whl.

File metadata

  • Download URL: llm_cockpit-1.0.0-py3-none-any.whl
  • Upload date:
  • Size: 872.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for llm_cockpit-1.0.0-py3-none-any.whl
Algorithm Hash digest
SHA256 7e05cae8ea54b53ec15ce7efd82ac3efc2b5f8bc5226c869fe9895b2badeebcf
MD5 2eda07713d3e3409714f95a959e626c7
BLAKE2b-256 b34a657bbb5fa08e3aeba7a695c6a43ec3bf6bdf34ec890bb669caf7c5703971

See more details on using hashes here.

Provenance

The following attestation bundles were made for llm_cockpit-1.0.0-py3-none-any.whl:

Publisher: release.yml on Bloxperts/llm-cockpit

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page