Skip to main content

Local LLM Council powered by Ollama

Project description

councillm

councillm is a lightweight, transparent LLM Council framework built on Ollama. It orchestrates multiple local language models into a structured decision-making pipeline inspired by Andrej Karpathy’s LLM Council concept — but designed for local-first, observable, and practical CLI usage.

This project focuses on correctness, transparency, and control, not theatrics.

License

PyPI Downloads

✨ Key Features

  • 🔁 Multi‑Model Reasoning (Generator → Critic → Chairman)
  • 🧠 Fast / Lite / Full execution modes
  • 🔍 Optional web‑search grounding
  • 👁️ Transparent execution logs — see each model work
  • 🖥️ Local‑only (no OpenAI / no cloud)
  • Fast install with uv

🏗️ Council Architecture

The system follows a strict, inspectable pipeline:

User Question
   ↓
[Generators]  → produce independent drafts
   ↓
[Critics]     → review & rank drafts (optional)
   ↓
[Chairman]   → synthesize final answer

Each stage is logged in real time so users can verify that the council is actually running.


📦 Installation (Fast with uv)

Prerequisites

  • Python 3.10+
  • Ollama installed and running
  • At least 2 local Ollama models pulled
ollama pull mistral
ollama pull llama3

Install with uv

uv pip install councillm

Or for development:

git clone https://github.com/yourname/councillm.git
cd councillm
uv pip install -e .

🚀 Quick Start

Run the CLI:

councillm

You will be prompted to configure the council once per session.


⚙️ Interactive Configuration

Instead of editing YAML files, councillm asks you directly:

Assign GENERATOR models (comma‑separated):
> mistral:latest, llama3:8b

Assign CRITIC models (comma‑separated):
> phi3:latest

Assign CHAIRMAN model:
> gemma3:1b

✔ No files are written ✔ No auto‑detection ✔ No hidden state


🧩 Execution Modes

After configuration, choose how the council runs:

1️⃣ Fast Mode

Chairman answers directly
  • Fastest
  • Lowest cost
  • Least robust

2️⃣ Lite Mode (Default)

Generator → Chairman
  • Balanced
  • Good for daily use

3️⃣ Full Mode

Multiple Generators → Critics → Chairman
  • Most reliable
  • Slowest
  • Maximum cross‑checking

🔍 Web Search Grounding

You can optionally enable web search:

Enable web search grounding? [y/N]: y

This uses DuckDuckGo search results to reduce hallucinations for factual queries.


🖥️ Example Session

$ councillm

======================================================================
        LLM COUNCIL — OLLAMA CONSOLE MODE
======================================================================
Type your question and press Enter.
Type 'exit' to quit.

Council ready (mode=full, search=True).

You: who won the 2020 F1 drivers championship?

[Stage 1] Generating responses
  • mistral:latest ✓
  • llama3:8b ✓

[Stage 2] Peer review
  • phi3:latest ✓

[Stage 3] Chairman synthesis
  • gemma3:1b ✓

Final Answer:
Lewis Hamilton won the 2020 Formula 1 Drivers' Championship.

🧪 Hallucination Mitigation Strategy

councillm reduces hallucinations by:

  • Multiple independent generations
  • Cross‑model critique
  • Chairman synthesis
  • Optional web grounding

⚠️ Still not perfect — this is risk reduction, not elimination.


❌ What councillm Is NOT

  • ❌ Not a chatbot UI
  • ❌ Not a prompt playground
  • ❌ Not a guarantee of truth
  • ❌ Not cloud‑based

This is a reasoning orchestrator, not a demo app.


📜 License

MIT License

You are free to use, modify, and distribute this project.


🤝 Contributing

Contributions are welcome if they:

  • Improve correctness
  • Reduce hallucinations
  • Increase transparency
  • Keep the system simple

🧠 Philosophy

"If you can’t observe it, you can’t trust it."

councillm exists to make local LLM reasoning inspectable, not magical.


Happy reasoning 🚀

Project details


Release history Release notifications | RSS feed

This version

1.0

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

councillm-1.0.tar.gz (16.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

councillm-1.0-py3-none-any.whl (16.1 kB view details)

Uploaded Python 3

File details

Details for the file councillm-1.0.tar.gz.

File metadata

  • Download URL: councillm-1.0.tar.gz
  • Upload date:
  • Size: 16.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for councillm-1.0.tar.gz
Algorithm Hash digest
SHA256 bb7fc261b2287aab6e5ffa70edffd1497f0c398b6702fc1abcee1a5b2cc8ecd0
MD5 e1a0fb84a9bd80ad1fc131146d66f703
BLAKE2b-256 632f0898dee93659bbef9df005125dfe7e2351bcad0dad1292931571e3b7bed1

See more details on using hashes here.

File details

Details for the file councillm-1.0-py3-none-any.whl.

File metadata

  • Download URL: councillm-1.0-py3-none-any.whl
  • Upload date:
  • Size: 16.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for councillm-1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 21c86ab5c0654d2a270733665803dec5f29d2a246f31bcf1aac932f356742b76
MD5 7cb11aae2d2c848842811ceda57c8827
BLAKE2b-256 d3dd98d73f7d48721864632cf21775066c00f45afe57af67d28cda9cf73cb3a6

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page