Skip to main content

Local-first reasoning pipeline wrapper for Ollama and LM Studio

Project description

🧠 MultiMind AI

A local-first web UI that adds sequential reasoning pipelines and parallel expert councils on top of small local models.

Python Version PyPI Version Local First Ollama Supported LM Studio Supported

Image


MultiMind AI bridges the gap between small local models and complex reasoning. It effortlessly auto-discovers endpoints like Ollama and LM Studio (OpenAI-compatible) and lets you deploy three reasoning architectures: a sequential Thinking Pipeline (Planning, Execution, Critique), a parallel Agent Council (Expert Advisors & Lead Judge), or a hierarchical Organisation Mode (CEO → Departments → Employees → Synthesis).


✨ Features

  • 🧠 Thinking Pipeline: Elevate smaller models with dedicated Planning, Execution, and Critique phases.
  • 🏛 Agent Council: Deploy a committee of expert models in parallel. Several 'Advisors' provide independent perspectives, synthesized by a Lead Judge into a single superior response.
  • 🏢 Organisation Mode: Run a hierarchical multi-agent workflow where a CEO decomposes the request, department heads delegate work to specialist roles, and the CEO synthesizes all outputs into one final answer.
  • 🔌 Zero-Config Auto-Discovery:
    • Automatically hooks into local Ollama endpoints (http://127.0.0.1:11434).
    • Supports optional discovery for LM Studio (http://127.0.0.1:1234).
  • 🎯 Precision Model Mapping: Assign distinct models to handle the different stages of thought or council roles.
  • 💬 Immersive UI: Enjoy a streaming timeline interface with collapsible "thought blocks" to keep your UI clean while the AI thinks.
  • 📝 Native Markdown & Math Support: Final outputs are rendered as HTML, with math equations typeset using a bundled local KaTeX build.
  • ⚡ Frictionless Setup: Purely in-memory settings. Zero .env setup required for your first run.

🚀 Quick Start

Get up and running in your local environment in seconds:

# 1. Install the package via pip
pip install multimind

# 2. Launch the application
multimind
🛠 Setting up for Development / Source Install
# 1. Clone the repository
git clone https://github.com/JitseLambrichts/MultiMind-AI.git
cd MultiMind-AI

# 2. Create a virtual environment
python3 -m venv .venv

# 3. Activate the virtual environment
source .venv/bin/activate  # On Windows: .venv\Scripts\activate

# 4. Install the package in editable mode
pip install -e .

# 5. Launch the application
multimind

Next: Open your browser and navigate to http://127.0.0.1:8000 🎉

🔌 Supported Backends

MultiMind AI works seamlessly with standard local APIs:

  • Ollama: Connects via /api/chat and /api/tags
  • OpenAI-Compatible Servers (e.g., LM Studio): Connects via /v1/chat/completions and /v1/models

If no provider is automatically detected, you can easily point the backend to your local OpenAI-compatible endpoint using the application's settings panel.

🧠 Thinking Pipeline (Sequential Reasoning)

The sequential reasoning pipeline elevates the capabilities of standard models by splitting inference into modular steps:

  1. Plan: Formulates a detailed technical roadmap to solve the user's request.
  2. Execute: Implements the primary solution based on the established plan.
  3. Critique (Hard Mode): Rigorously audits the implementation for errors or omissions, delivering a refined, superior final answer.

🏛 Agent Council (Parallel Expert Consensus)

For complex tasks requiring multiple perspectives, MultiMind AI offers the Agent Council. This architecture emphasizes parallel expertise over sequential steps:

  1. Parallel Advisors: Multiple models process the user's request independently, providing diverse expert viewpoints.
  2. Diverse Perspectives: Each advisor follows expert-level system prompts to ensure accurate, independent technical analysis.
  3. The Judge: A final 'Lead Synthesizer' model reviews all advisor outputs, cross-examines their findings, resolves conflicts, and merges the best elements into a single definitive response.

🏢 Organisation Mode (Hierarchical Multi-Agent Workflow)

For tasks that benefit from structured delegation, MultiMind AI includes Organisation Mode. Instead of parallel peers only, this mode simulates an org chart with explicit delegation layers:

  1. CEO Planning: A CEO agent analyzes the user request and splits it into department-level sub-tasks.
  2. Department Delegation: Each department head converts its sub-task into role-specific assignments.
  3. Employee Execution: Specialist employee agents execute their assigned tasks and stream their outputs.
  4. CEO Synthesis: The CEO consolidates all department/employee results into a single cohesive final response.

In the UI, this appears as an interactive organisation chart with expandable nodes and streaming outputs per role, while still ending with one final answer block.

⚙️ Configuration: Organisation mode uses one selected model for all agents in the hierarchy (configurable via the Organisation settings panel).

📝 Note: Chat history is intentionally in-memory only for the current MVP.

📊 Benchmarks

We evaluated the performance of MultiMind AI's reasoning pipeline using a subset of 20 questions from the GSM8K dataset. The results demonstrate a clear improvement in model accuracy when utilizing the different reasoning modes.

Image

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

multimind-0.1.8.tar.gz (1.5 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

multimind-0.1.8-py3-none-any.whl (1.6 MB view details)

Uploaded Python 3

File details

Details for the file multimind-0.1.8.tar.gz.

File metadata

  • Download URL: multimind-0.1.8.tar.gz
  • Upload date:
  • Size: 1.5 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.1

File hashes

Hashes for multimind-0.1.8.tar.gz
Algorithm Hash digest
SHA256 7569fee67eb089ef4a6c4e228e918d5b078f9c4ba4dbd5fd027ae11998c14e29
MD5 6dc9c2408d3ece6d35f5cc8b481275fc
BLAKE2b-256 fa8823c8ab85eb74c4a59f29029aac86ea44d17d4b3c29b615e92353a4b0c3e4

See more details on using hashes here.

File details

Details for the file multimind-0.1.8-py3-none-any.whl.

File metadata

  • Download URL: multimind-0.1.8-py3-none-any.whl
  • Upload date:
  • Size: 1.6 MB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.1

File hashes

Hashes for multimind-0.1.8-py3-none-any.whl
Algorithm Hash digest
SHA256 73f5cdfa44eae37666a954ab2b47611b8693df8cc231396eaea32969d8bee535
MD5 132d4536089f861e248fbb40fb76982b
BLAKE2b-256 2082bee52d7a4f97084e3580ea8fabdc6c66c52280912159d3c951cc04831e3b

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page