Skip to main content

Open Swarm: Orchestrating AI Agent Swarms with Django

Project description

Open Swarm

Project Logo

Open Swarm is a Python framework for creating, managing, and deploying autonomous agent swarms. It leverages the openai-agents library for core agent functionality and provides a structured way to build complex, multi-agent workflows using Blueprints.

Open Swarm can be used in two primary ways:

  1. As a CLI Utility (swarm-cli): Manage, run, and install blueprints directly on your local machine. Ideal for personal use, testing, and creating standalone agent tools. (Recommended installation: PyPI)
  2. As an API Service (swarm-api): Deploy a web server that exposes your blueprints via an OpenAI-compatible REST API. Ideal for integrations, web UIs, and shared access. (Recommended deployment: Docker)

Core Concepts

  • Agents: Individual AI units performing specific tasks, powered by LLMs (like GPT-4, Claude, etc.). Built using the openai-agents SDK.
  • Blueprints: Python classes (BlueprintBase subclasses) defining a swarm's structure, agents, coordination logic, and external dependencies (like required environment variables or MCP servers). They act as reusable templates for specific tasks (e.g., code generation, research, data analysis).
  • MCP (Model Context Protocol) Servers: Optional external processes providing specialized capabilities (tools) to agents, such as filesystem access, web browsing, database interaction, or interacting with specific APIs (Slack, Monday.com, etc.). Agents interact with MCP servers via a standardized communication protocol.
  • Configuration (swarm_config.json): A central JSON file defining available LLM profiles (API keys, models) and configurations for MCP servers. Typically managed via swarm-cli in ~/.config/swarm/.
  • swarm-cli: A command-line tool for managing blueprints (adding, listing, running, installing) and the swarm_config.json file. Uses XDG directories for storing blueprints (~/.local/share/swarm/blueprints/) and configuration (~/.config/swarm/).
  • swarm-api: A launcher for the Django/DRF backend that exposes installed blueprints via an OpenAI-compatible REST API (/v1/models, /v1/chat/completions).

Environment Variables

Open Swarm and its blueprints use a variety of environment variables for configuration, security, and integration with external services. Set these in your shell, .env file, Docker environment, or deployment platform as appropriate.

Core Framework Environment Variables

Variable Description Default / Required
OPENAI_API_KEY API key for OpenAI LLMs (used by agents and blueprints) Required for OpenAI usage
SWARM_API_KEY API key for securing API endpoints (swarm-api) Optional (recommended)
LITELLM_BASE_URL Override base URL for LiteLLM/OpenAI-compatible endpoints Optional
LITELLM_API_KEY API key for LiteLLM endpoints Optional
SWARM_CONFIG_PATH Path to the main Swarm config file (swarm_config.json) ../swarm_config.json
BLUEPRINT_DIRECTORY Directory containing blueprint files src/swarm/blueprints
DJANGO_SECRET_KEY Django secret key (for API mode) Auto-generated/dev default
DJANGO_DEBUG Enable Django debug mode True
DJANGO_ALLOWED_HOSTS Comma-separated allowed hosts for Django API localhost,127.0.0.1
API_AUTH_TOKEN Token for authenticating API requests Optional
DJANGO_LOG_LEVEL Log level for Django app INFO
SWARM_LOG_LEVEL Log level for Swarm app DEBUG
REDIS_HOST Host for Redis (if used) localhost
REDIS_PORT Port for Redis (if used) 6379
DJANGO_CSRF_TRUSTED_ORIGINS Comma-separated trusted origins for CSRF protection http://localhost:8000,...
ENABLE_ADMIN Enable admin web interface false
ENABLE_API_AUTH Require API authentication true

Blueprint/Tool-Specific Variables

  • Some blueprints and MCP tools may require additional env vars (e.g., Google API keys, Slack tokens, etc.).
  • Refer to the blueprint's docstring or config for details.

Usage Example

export OPENAI_API_KEY="sk-..."
export SWARM_API_KEY="..."
export LITELLM_BASE_URL="https://open-litellm.fly.dev/v1"
# ... set other variables as needed

Toolbox Functionality

Open Swarm ships with a growing toolbox of agent and blueprint utilities. All features listed below have robust, passing tests unless marked as WIP (Work In Progress).

Task Scheduler Toolbox

  • Schedule jobs with at:
    • Schedule a shell script or command to run at a specific time (uses the system at command).
    • Test Status: Passing
  • List scheduled at jobs:
    • List all jobs currently scheduled with at.
    • Test Status: Passing
  • Remove at jobs:
    • Remove a scheduled job by its job ID.
    • Test Status: Passing
  • Schedule jobs with cron:
    • Schedule recurring jobs using cron expressions (uses the system crontab).
    • Test Status: Passing
  • List scheduled cron jobs:
    • List all jobs currently scheduled with crontab.
    • Test Status: Passing
  • Remove cron jobs:
    • Remove a scheduled cron job by its job ID.
    • Test Status: Passing

Slash Command Framework

  • Global slash command registry:
    • Blueprints can register and use slash commands (e.g., /help, /agent, /model).
    • Built-in demo commands: /help, /agent, /model.
    • Test Status: Passing
  • Blueprint Integration:
    • Blueprints can access the global registry and add their own commands.
    • Test Status: Passing

Usage Example (Slash Commands)

from swarm.extensions.blueprint.slash_commands import slash_command_registry

@slash_command_registry.register('/hello')
def hello_command(args):
    return f"Hello, {args}!"

Usage Example (Task Scheduler)

from swarm.extensions.task_scheduler_toolbox import schedule_at_job, list_at_jobs, remove_at_job

job_id = schedule_at_job('/path/to/script.sh', run_time='now + 5 minutes')
jobs = list_at_jobs()
remove_at_job(job_id)

Developer Notes

  • System dependencies are mocked in tests for CI and portability.
  • Any toolbox feature not listed as Passing above is considered WIP and may not be stable.
  • Contributions and feedback are welcome!

Quickstart 1: Using swarm-cli Locally (via PyPI)

This is the recommended way to use swarm-cli for managing and running blueprints on your local machine.

Prerequisites:

  • Python 3.10+
  • pip (Python package installer)

Steps:

  1. Install open-swarm from PyPI:

    pip install open-swarm
    

    (Using a virtual environment is recommended: python -m venv .venv && source .venv/bin/activate)

  2. Initial Configuration (First Run):

    • The first time you run a swarm-cli command that requires configuration (like run or config), it will automatically create a default swarm_config.json at ~/.config/swarm/swarm_config.json if one doesn't exist.
    • You must set the required environment variables (like OPENAI_API_KEY) in your shell for the configuration to work. Create a .env file in your working directory or export them:
      export OPENAI_API_KEY="sk-..."
      # Add other keys as needed (GROQ_API_KEY, etc.)
      
    • You can customize the configuration further using swarm-cli config commands (see USERGUIDE.md).
  3. Add a Blueprint:

    • Download or create a blueprint file (e.g., my_blueprint.py). Example blueprints are available in the project repository.
    • Add it using swarm-cli:
      # Example: Adding a downloaded blueprint file
      swarm-cli add ./path/to/downloaded/blueprint_echocraft.py
      
      # Example: Adding a directory containing a blueprint
      swarm-cli add ./my_custom_blueprints/agent_smith --name agent_smith
      
  4. Run the Blueprint:

    • Single Instruction:
      swarm-cli run echocraft --instruction "Hello from CLI!"
      
    • Interactive Mode:
      swarm-cli run echocraft
      # Now you can chat with the blueprint interactively
      
  5. (Optional) Install as Command:

    swarm-cli install echocraft
    # Now run (ensure ~/.local/share/swarm/bin is in your PATH):
    echocraft --instruction "I am a command now!"
    

Quickstart 2: Deploying swarm-api Service (via Docker)

This section covers deploying the API service using Docker.

Option A: Docker Compose (Recommended for Flexibility)

This method uses docker-compose.yaml and is best if you need to customize volumes, environment variables easily, or manage related services (like Redis).

Prerequisites:

Steps:

  1. Clone the Repository: (Needed for docker-compose.yaml and config files)

    git clone https://github.com/matthewhand/open-swarm.git
    cd open-swarm
    
  2. Configure Environment:

    • Copy cp .env.example .env and edit .env with your API keys (e.g., OPENAI_API_KEY, SWARM_API_KEY).
  3. Prepare Blueprints & Config:

    • Place blueprints in ./blueprints.
    • Ensure ./swarm_config.json exists and is configured.
  4. Configure Overrides (Optional):

    • Copy cp docker-compose.override.yaml.example docker-compose.override.yaml.
    • Edit the override file to mount additional volumes, change ports, etc.
  5. Start the Service:

    docker compose up -d
    
  6. Verify API: (Default port 8000)

    • Models: curl http://localhost:8000/v1/models
    • Chat: curl http://localhost:8000/v1/chat/completions -H "Content-Type: application/json" -d '{"model": "echocraft", ...}' (Add -H "Authorization: Bearer <key>" if needed).

Option B: Direct docker run (Simpler for Single Container)

This method runs the pre-built image directly from Docker Hub. Good for quick tests or simple deployments without cloning the repo. Customization requires careful use of -v (volume) and -e (environment) flags.

Prerequisites:

Steps:

  1. Prepare Local Files (If Customizing):

    • Create a directory for your blueprints (e.g., ~/my_swarm_blueprints).
    • Create your swarm_config.json file locally (e.g., ~/my_swarm_config.json).
    • Create a .env file locally (e.g., ~/swarm.env) with your API keys (OPENAI_API_KEY, SWARM_API_KEY, etc.).
  2. Run the Container:

    docker run -d \
      --name open-swarm-api \
      -p 8000:8000 \
      --env-file ~/swarm.env \
      -v ~/my_swarm_blueprints:/app/blueprints:ro \
      -v ~/my_swarm_config.json:/app/swarm_config.json:ro \
      -v open_swarm_db:/app/db.sqlite3 \
      --restart unless-stopped \
      mhand79/open-swarm:latest
    
    • -d: Run detached (in background).
    • --name: Assign a name to the container.
    • -p 8000:8000: Map host port 8000 to container port 8000 (adjust if needed).
    • --env-file: Load environment variables from your local file.
    • -v ...:/app/blueprints:ro: Mount your local blueprints directory (read-only). Required if you want to use custom blueprints.
    • -v ...:/app/swarm_config.json:ro: Mount your local config file (read-only). Required for custom LLM/MCP settings.
    • -v open_swarm_db:/app/db.sqlite3: Use a named Docker volume for the database to persist data.
    • --restart unless-stopped: Automatically restart the container unless manually stopped.
    • mhand79/open-swarm:latest: The image name on Docker Hub.
  3. Verify API: (Same as Docker Compose)

    • Models: curl http://localhost:8000/v1/models
    • Chat: curl http://localhost:8000/v1/chat/completions ... (Add -H "Authorization: Bearer <key>" if needed).

Usage Modes Summary

  • swarm-api (via Docker or manage.py runserver): Exposes blueprints as an OpenAI-compatible REST API. Ideal for integrations. Requires SWARM_API_KEY for security in non-local deployments.
  • swarm-cli run (via PyPI install): Executes managed blueprints locally, either with a single instruction or in interactive chat mode. Good for testing and local tasks.
  • swarm-cli install (via PyPI install): Creates standalone command-line executables from managed blueprints.
  • Direct Python Execution (via Git clone): Running uv run python <blueprint_file.py> is mainly for development and testing individual files.

Further Documentation

This README provides a high-level overview and quickstart guides. For more detailed information, please refer to:

  • User Guide (USERGUIDE.md): Detailed instructions on using swarm-cli commands for managing blueprints and configuration locally.
  • Development Guide (DEVELOPMENT.md): Information for contributors and developers, including architecture details, testing strategies, project layout, API details, and advanced topics.
  • Example Blueprints (src/swarm/blueprints/README.md): A list and description of the example blueprints included with the framework, showcasing various features and integration patterns.

Contributing

Contributions are welcome! Please refer to the CONTRIBUTING.md file (if available) or open an issue/pull request on the repository.


License

Open Swarm is provided under the MIT License. Refer to the LICENSE file for full details.


Acknowledgements

This project builds upon concepts and code from the openai-agents library and potentially other open-source projects. Specific acknowledgements can be found in DEVELOPMENT.md or individual source files.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

open_swarm-0.1.1744952955.tar.gz (225.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

open_swarm-0.1.1744952955-py3-none-any.whl (308.5 kB view details)

Uploaded Python 3

File details

Details for the file open_swarm-0.1.1744952955.tar.gz.

File metadata

  • Download URL: open_swarm-0.1.1744952955.tar.gz
  • Upload date:
  • Size: 225.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.12.9

File hashes

Hashes for open_swarm-0.1.1744952955.tar.gz
Algorithm Hash digest
SHA256 fc71bf7739b7ae6c6bd71677d8bb91d87fdf1f602378be5ef0be52ec66f5de7e
MD5 7d952a540962a3b64e438c9558836c71
BLAKE2b-256 e40f6adb3d0ad90e8298fd0171e252da95cb25b9e2d26a8ef39e0968cf92f5db

See more details on using hashes here.

File details

Details for the file open_swarm-0.1.1744952955-py3-none-any.whl.

File metadata

File hashes

Hashes for open_swarm-0.1.1744952955-py3-none-any.whl
Algorithm Hash digest
SHA256 a023f4648d04f862c7b7ef2629e19076c98e62d7eddf8c296d472ab853000c4e
MD5 cfa6f4a9f6cf51c0d58bab59632ac3ee
BLAKE2b-256 41a23d7a9b5fd5f28b4a09623e0e062667363e11485f32e8589a1e97ddd93e08

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page