Skip to main content

Local-first reasoning pipeline wrapper for Ollama and LM Studio

Project description

🧠 MultiMind AI

A local-first web UI that adds a reasoning pipeline on top of small local models.

Python Version Local First Ollama Supported LM Studio Supported


MultiMind AI acts as an intelligent reasoning pipeline for your local AI models. It effortlessly auto-discovers endpoints like Ollama and LM Studio (OpenAI-compatible) and lets you orchestrate dedicated models for different logical phases: Planning, Execution, and Critique.


✨ Features

  • 🧠 Adaptive Reasoning Modes: Toggle between Off, Medium, and Hard modes to dictate the depth of the model's reflection.
  • 🔌 Zero-Config Auto-Discovery:
    • Automatically hooks into local Ollama endpoints (http://127.0.0.1:11434).
    • Supports optional discovery for LM Studio (http://127.0.0.1:1234).
  • 🎯 Precision Model Mapping: Assign distinct models to handle the different stages of thought (plan, execute, and critique).
  • 💬 Immersive UI: Enjoy a streaming timeline interface with collapsible "thought blocks" to keep your UI clean while the AI thinks.
  • 📝 Native Markdown & Math Support:
    • Final outputs are beautifully rendered as HTML in the chat view.
    • Inline and block math equations are flawlessly typeset using a bundled local KaTeX build.
  • ⚡ Frictionless Setup: Purely in-memory settings. Zero .env setup required for your first run.

🚀 Quick Start

Get up and running in your local environment in seconds:

# 1. Create a virtual environment
python3 -m venv .venv

# 2. Activate the virtual environment
# On macOS / Linux:
source .venv/bin/activate

# On Windows:
.venv\Scripts\activate

# 3. Install the package
pip install -e .

# 4. Launch the application
multimind AI

Next: Open your browser and navigate to http://127.0.0.1:8000 🎉

🔌 Supported Backends

MultiMind AI works seamlessly with standard local APIs:

  • Ollama: Connects via /api/chat and /api/tags
  • OpenAI-Compatible Servers (e.g., LM Studio): Connects via /v1/chat/completions and /v1/models

If no provider is automatically detected, you can easily point the backend to your local OpenAI-compatible endpoint using the application's settings panel.

💡 How It Works

MultiMind AI splits inference into modular steps, elevating the capabilities of standard models:

  1. Plan: Formulates a structured approach to the prompt.
  2. Execute: Generates the primary response.
  3. Critique (Hard Mode): Evaluates the execution pass as a rough draft and streams refined, critiqued output as the final answer.

📝 Note: Chat history is intentionally in-memory only for the current MVP.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

multimind-0.1.0.tar.gz (1.3 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

multimind-0.1.0-py3-none-any.whl (1.4 MB view details)

Uploaded Python 3

File details

Details for the file multimind-0.1.0.tar.gz.

File metadata

  • Download URL: multimind-0.1.0.tar.gz
  • Upload date:
  • Size: 1.3 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.1

File hashes

Hashes for multimind-0.1.0.tar.gz
Algorithm Hash digest
SHA256 b7d7764fc45079649a5cce5ae46fe561ed80717230b70ce3f1d65f3425a99187
MD5 9ee29dc3f1bd1b84bc903fddf80a6455
BLAKE2b-256 9ca705789d72c603ab933634b186bad945a53f60892977ae447d0fe939fd6bee

See more details on using hashes here.

File details

Details for the file multimind-0.1.0-py3-none-any.whl.

File metadata

  • Download URL: multimind-0.1.0-py3-none-any.whl
  • Upload date:
  • Size: 1.4 MB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.1

File hashes

Hashes for multimind-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 ded7d20bbd4ad24a1d151f15b0ac0476d5c0342bf0f02ba3d96b8249d83e3603
MD5 9a9a08b6b3d0baf2caf815dd2250e117
BLAKE2b-256 202c2852d0c795de7c18b75a1e1bd62c73c7fa6bad13cd47f1a591cb08c69886

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page