Skip to main content

A flexible, self-hosted RAG chatbot framework for containerized deployments.

Project description

Kondoo 🦙

Kondoo is not just a chatbot; it is a framework for building autonomous digital minds. Its name is inspired by the word “condominium,” a system of independent dwellings that share the same structure. Similarly, Kondoo allows multiple bots to operate independently, each with its own personality and knowledge base, but sharing the same robust, containerized framework.

This project was born with a “self-hosted first” philosophy, giving you complete control over your data and the models you use, from a local tinyllama to cloud APIs such as Gemini.

Kondoo: Your knowledge, your rules, your assistants.


🚀 Key Features

  • Framework Agnostic: Not tied to a specific provider. Use an ANSWER_LLM_PROVIDER to choose your answer engine (Gemini, OpenAI, Ollama) and a KNOWLEDGE_PROVIDER for your embeddings (Ollama, local, OpenAI).
  • Containerized by Design: Built on Podman and compose, ensuring maximum portability and clean, repeatable deployment.
  • Self-Hosted First: Designed to run 100% locally, using Ollama for both embeddings and response generation, giving you full control and privacy.
  • Extensible: The src/ structure makes it an installable Python package, ready to be imported into larger projects.
  • Decoupled Identity: Separates the bot's identity (persona.yaml) from its behavioral rules (behavior.txt), allowing for scalable management of multiple bots with standardized service quality.

🏛️ Project Structure

Kondoo is structured as a Python framework, separating reusable code from implementation examples:

  • src/kondoo/: The source code for the kondoo framework (installable via pip).
  • example/example_bot/: A complete and functional example bot that shows how to use the framework. This is your starting point.
  • pyproject.toml: Defines the project and all its dependencies.
  • .env.example: A universal template with all available environment variables.

⚡ Quickstart Guide

Try Kondoo in 5 minutes using the sample bot.

1. Prerequisites

  • Podman and podman-compose.
  • Python 3.9+
  • Your own Ollama service (local or remote) or an API Key (e.g., Google Gemini).
  • SynapsIA to create the knowledge base.

2. Clone the Repository

git clone https://github.com/sysadminctl-services/kondoo.git
cd kondoo

3. Set Up the Example Bot

Copy the configuration template to the example bot directory:

cp .env.example example/example_bot/.env

Edit the .env file and fill in the variables. For a 100% local test with Ollama:

# example/example_bot/.env
ANSWER_LLM_PROVIDER=ollama_compatible
KNOWLEDGE_PROVIDER=ollama

LLM_MODEL_NAME="tinyllama"
LLM_BASE_URL="http://host.containers.internal:11434/v1"

EMBEDDING_MODEL_NAME="mxbai-embed-large"
OLLAMA_BASE_URL="http://host.containers.internal:11434"

BOT_PERSONA_FILE=/app/persona.yaml
BOT_BEHAVIOR_FILE=/app/behavior.txt

4. Create the Knowledge Base

Create the directories for the documents and the knowledge base:

cd example/example_bot
mkdir -p docs knowledge
  1. Define Identity: Edit persona.yaml to define the bot's name and role.

  2. Define Behavior: Edit behavior.txt to set the interaction rules.

  3. Create Documents: Add your source files to the docs/ folder.

echo "Kondoo is a RAG chatbot framework created by sysadminctl.services." > docs/info.txt
  1. Ingest Knowledge: Use the installed synapsia command:
synapsia --docs ./docs/ --knowledge ./knowledge/

5. Launch the Container

Return to the bot directory and run podman-compose:

# While in example/example_bot/
podman-compose up --build

6. Test the Bot

Open a new terminal and send a query using curl:

curl -X POST \
  -H "Content-Type: application/json" \
  -d '{"query": "What is Kondoo?"}' \
  http://localhost:5000/query

You should receive a JSON response generated by your local tinyllama. 114:

7. Chat Mode (History)

To maintain a conversation history, add a session_id to your request:

curl -X POST \
  -H "Content-Type: application/json" \
  -d '{"query": "My name is Luis", "session_id": "user-123"}' \
  http://localhost:5000/query

The bot will remember context for that specific session ID (stored in RAM, resets on restart).

⚙️ Configuration (.env)

All configuration variables are documented in the .env.example file. Variables are loaded from .env in your bot's directory (e.g., example/example_bot/.env).

1. Provider Selection

These variables act as "switches" to choose which services to use.

  • ANSWER_LLM_PROVIDER: Choose your response (LLM) engine.
    • gemini: (Cloud) Google Gemini (requires LLM_API_KEY).
    • openai: (Cloud) OpenAI (requires LLM_API_KEY).
    • ollama_compatible: (Self-Hosted) Any OpenAI-compatible API, like Ollama (requires LLM_BASE_URL and LLM_MODEL_NAME).
  • KNOWLEDGE_PROVIDER: Choose your embeddings (knowledge) engine.
    • ollama: (Self-Hosted) Use an Ollama service (requires OLLAMA_BASE_URL and EMBEDDING_MODEL_NAME).
    • local: (Local) Use a HuggingFace model on the CPU/GPU (requires EMBEDDING_MODEL_NAME).
    • openai: (Cloud) Use OpenAI's embeddings API (requires LLM_API_KEY).

2. Provider-Specific Settings

These are the "control knobs" required by the providers you selected above.

Answer Engine (LLM) Settings

  • LLM_API_KEY:
    • Required by: gemini, openai.
    • Description: Your secret API key for the chosen cloud service.
  • LLM_MODEL_NAME:
    • Required by: gemini, openai, ollama_compatible.
    • Description: The specific model name to use for generating answers.
    • Examples: models/gemini-1.5-flash, gpt-4o, tinyllama.
  • LLM_BASE_URL:
    • Required by: ollama_compatible.
    • Description: The full base URL of your self-hosted LLM's OpenAI-compatible API.
    • Example (Ollama): http://host.containers.internal:11434/v1

Knowledge (Embedding) Settings

  • EMBEDDING_MODEL_NAME:
    • Required by: ollama, local, openai.
    • Description: The specific model name to use for embeddings.
    • Examples: mxbai-embed-large, nomic-embed-text.
  • OLLAMA_BASE_URL:
    • Required by: ollama (provider).
    • Description: The base URL of your Ollama service (the non-/v1 endpoint).
    • Example: http://host.containers.internal:11434


3. Fine Tuning

Customize the bot's creativity and retrieval precision.

  • LLM_TEMPERATURE:
    • Description: Controls the randomness of the model. 0.0 is deterministic (good for data extraction), 1.0 is creative.
    • Default: 0.1
  • RAG_TOP_K:
    • Description: Number of text chunks to retrieve from the knowledge base for each query.
    • Default: 2

4. Bot Configuration

These variables control the bot's identity and data paths.

  • BOT_PERSONA_FILE:
    • Description: Path to the YAML file defining the specific identity (Name, Role, Tone).
    • Default: /app/persona.yaml.
  • BOT_BEHAVIOR_FILE:
    • Description: Path to the TXT or Markdown file defining global interaction rules.
    • Default: /app/behavior.txt.
  • KNOWLEDGE_DIR:
    • Description: The path where the bot will load its knowledge base from.
    • Default: /app/knowledge.

⚖️ License

This project is licensed under the MIT License. See the LICENSE file for more details.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

kondoo-0.1.10.tar.gz (12.9 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

kondoo-0.1.10-py3-none-any.whl (10.1 kB view details)

Uploaded Python 3

File details

Details for the file kondoo-0.1.10.tar.gz.

File metadata

  • Download URL: kondoo-0.1.10.tar.gz
  • Upload date:
  • Size: 12.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for kondoo-0.1.10.tar.gz
Algorithm Hash digest
SHA256 80429aef52d3f6a1ee12090d10d50f095fa601ce4b3bcb10b067dcbeac0ce085
MD5 a077adbb07b9b8d7590312896189353e
BLAKE2b-256 9c356055f726ff2724b873c9453e66809338449e8fef0fa6e0fbcf96e94125b5

See more details on using hashes here.

Provenance

The following attestation bundles were made for kondoo-0.1.10.tar.gz:

Publisher: publish-to-pypi.yml on sysadminctl-services/kondoo

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file kondoo-0.1.10-py3-none-any.whl.

File metadata

  • Download URL: kondoo-0.1.10-py3-none-any.whl
  • Upload date:
  • Size: 10.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for kondoo-0.1.10-py3-none-any.whl
Algorithm Hash digest
SHA256 5ea4a67aff8242a40ec5c47d849d64790943aa13ae234d1264cec32c0da0e34c
MD5 0194c3b42f9f9894e9070e6e996195bc
BLAKE2b-256 747a10c6b383ceb7344ddfcaba606d1a35930d386715120ef4e3075d4108bb61

See more details on using hashes here.

Provenance

The following attestation bundles were made for kondoo-0.1.10-py3-none-any.whl:

Publisher: publish-to-pypi.yml on sysadminctl-services/kondoo

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page