Local-first reasoning pipeline wrapper for Ollama and LM Studio
Project description
🧠 MultiMind AI
A local-first web UI that adds sequential reasoning pipelines and parallel expert councils on top of small local models.
MultiMind AI bridges the gap between small local models and complex reasoning. It effortlessly auto-discovers endpoints like Ollama and LM Studio (OpenAI-compatible) and lets you deploy three reasoning architectures: a sequential Thinking Pipeline (Planning, Execution, Critique), a parallel Agent Council (Expert Advisors & Lead Judge), or a hierarchical Organisation Mode (CEO → Departments → Employees → Synthesis).
✨ Features
- 🧠 Thinking Pipeline: Elevate smaller models with dedicated Planning, Execution, and Critique phases.
- 🏛 Agent Council: Deploy a committee of expert models in parallel. Several 'Advisors' provide independent perspectives, synthesized by a Lead Judge into a single superior response.
- 🏢 Organisation Mode: Run a hierarchical multi-agent workflow where a CEO decomposes the request, department heads delegate work to specialist roles, and the CEO synthesizes all outputs into one final answer.
- 🔌 Zero-Config Auto-Discovery:
- Automatically hooks into local Ollama endpoints (
http://127.0.0.1:11434). - Supports optional discovery for LM Studio (
http://127.0.0.1:1234).
- Automatically hooks into local Ollama endpoints (
- 🎯 Precision Model Mapping: Assign distinct models to handle the different stages of thought or council roles.
- 💬 Immersive UI: Enjoy a streaming timeline interface with collapsible "thought blocks" to keep your UI clean while the AI thinks.
- 📝 Native Markdown & Math Support: Final outputs are rendered as HTML, with math equations typeset using a bundled local KaTeX build.
- ⚡ Frictionless Setup: Purely in-memory settings. Zero
.envsetup required for your first run.
🚀 Quick Start
Get up and running in your local environment in seconds:
# 1. Install the package via pip
pip install multimind
# 2. Launch the application
multimind
🛠 Setting up for Development / Source Install
# 1. Clone the repository
git clone https://github.com/JitseLambrichts/MultiMind-AI.git
cd MultiMind-AI
# 2. Create a virtual environment
python3 -m venv .venv
# 3. Activate the virtual environment
source .venv/bin/activate # On Windows: .venv\Scripts\activate
# 4. Install the package in editable mode
pip install -e .
# 5. Launch the application
multimind
Next: Open your browser and navigate to http://127.0.0.1:8000 🎉
🔌 Supported Backends
MultiMind AI works seamlessly with standard local APIs:
- Ollama: Connects via
/api/chatand/api/tags - OpenAI-Compatible Servers (e.g., LM Studio): Connects via
/v1/chat/completionsand/v1/models
If no provider is automatically detected, you can easily point the backend to your local OpenAI-compatible endpoint using the application's settings panel.
🧠 Thinking Pipeline (Sequential Reasoning)
The sequential reasoning pipeline elevates the capabilities of standard models by splitting inference into modular steps:
- Plan: Formulates a detailed technical roadmap to solve the user's request.
- Execute: Implements the primary solution based on the established plan.
- Critique (Hard Mode): Rigorously audits the implementation for errors or omissions, delivering a refined, superior final answer.
🏛 Agent Council (Parallel Expert Consensus)
For complex tasks requiring multiple perspectives, MultiMind AI offers the Agent Council. This architecture emphasizes parallel expertise over sequential steps:
- Parallel Advisors: Multiple models process the user's request independently, providing diverse expert viewpoints.
- Diverse Perspectives: Each advisor follows expert-level system prompts to ensure accurate, independent technical analysis.
- The Judge: A final 'Lead Synthesizer' model reviews all advisor outputs, cross-examines their findings, resolves conflicts, and merges the best elements into a single definitive response.
🏢 Organisation Mode (Hierarchical Multi-Agent Workflow)
For tasks that benefit from structured delegation, MultiMind AI includes Organisation Mode. Instead of parallel peers only, this mode simulates an org chart with explicit delegation layers:
- CEO Planning: A CEO agent analyzes the user request and splits it into department-level sub-tasks.
- Department Delegation: Each department head converts its sub-task into role-specific assignments.
- Employee Execution: Specialist employee agents execute their assigned tasks and stream their outputs.
- CEO Synthesis: The CEO consolidates all department/employee results into a single cohesive final response.
In the UI, this appears as an interactive organisation chart with expandable nodes and streaming outputs per role, while still ending with one final answer block.
⚙️ Configuration: Organisation mode uses one selected model for all agents in the hierarchy (configurable via the Organisation settings panel).
📝 Note: Chat history is intentionally in-memory only for the current MVP.
📊 Benchmarks
We evaluated the performance of MultiMind AI's reasoning pipeline using a subset of 20 questions from the GSM8K dataset. The results demonstrate a clear improvement in model accuracy when utilizing the different reasoning modes.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file multimind-0.2.0.tar.gz.
File metadata
- Download URL: multimind-0.2.0.tar.gz
- Upload date:
- Size: 1.5 MB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.14.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
90bc74557ef033d6fd9780ed283a7b8262564939fc2c7c248ca7db0f923a933e
|
|
| MD5 |
45fd3760a0fc989a37a9e70376791ea7
|
|
| BLAKE2b-256 |
2fcc1295dec283c87aad1f04e9059061b3415eabb4d6ae00c37d4dc364844c66
|
File details
Details for the file multimind-0.2.0-py3-none-any.whl.
File metadata
- Download URL: multimind-0.2.0-py3-none-any.whl
- Upload date:
- Size: 1.6 MB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.14.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
2c1e9afcad7d041b69f88d22439383cc8bfcd679ce2d27c1f6d3fd20e82317ab
|
|
| MD5 |
99501e87205b9c3bac11eda6354d3795
|
|
| BLAKE2b-256 |
3c7c413ccfb0365a430c40764f1151540a5fa6cc6dc734c0dc4795d49b98f2a7
|