Skip to main content

Debuggable runtime for AI agent pipelines

Project description


Binex

Open-source visual orchestrator for AI agent workflows
Build, run, debug, and replay multi-agent pipelines — 100% locally.

PyPI Python License CI Docs Stars

Demo · Install · Web UI · Features · Docs · Issues



Demo

Full workflow: drag & drop nodes, configure models and prompts, run with human input, see results, debug, trace, and lineage — all in the browser.

Binex Demo
Click to watch full demo video

(back to top)


What is Binex?

Binex is an open-source, fully local runtime for AI agent workflows. No cloud. No telemetry. No vendor lock-in.

pip install binex
binex ui

That's it. Browser opens. You're building AI workflows.

Why Binex?

  • 100% local — your data never leaves your machine
  • 100% open source — MIT licensed, audit every line
  • Zero telemetry — no tracking, no analytics, no surprises
  • Full debuggability — every input, output, prompt, and cost is visible
  • Any model — OpenAI, Anthropic, Google, Ollama, OpenRouter, DeepSeek, and 40+ more via LiteLLM

(back to top)


Installation

pip install binex

With extras:

pip install binex[langchain]    # LangChain Runnables
pip install binex[crewai]       # CrewAI Crews
pip install binex[autogen]      # AutoGen Teams
pip install binex[telemetry]    # OpenTelemetry tracing
pip install binex[rich]         # Rich colored CLI output

(back to top)


Web UI

Launch the visual workflow editor:

binex ui

Visual Drag & Drop Editor

Build workflows visually — drag nodes from the palette, connect them, configure models and prompts inline.

6 node types: LLM Agent, Local Script, Human Input, Human Approve, Human Output, A2A Agent

  • 20+ preset models including 8 free OpenRouter models
  • Built-in prompt library (Planner, Researcher, Analyzer, Writer, Reviewer, Summarizer)
  • Switch between Visual and YAML modes — changes sync both ways
  • Real-time cost estimation as you build
  • Custom model input — use any litellm-compatible model

18 Pages — Full CLI Parity

Category Pages
Workflows Browse, Visual Editor, Scaffold Wizard
Runs Dashboard, RunLive (SSE), RunDetail
Analysis Debug (input/output artifacts), Trace (Gantt timeline), Diagnose (root-cause), Lineage (artifact graph)
Comparison Diff (side-by-side), Bisect (find divergence)
Costs Cost Dashboard (charts), Budget Management
System Doctor (health), Plugins, Gateway, Export

Replay

Debug any node → click Replay → swap the model or prompt → re-run just that node. No re-running the entire pipeline.

(back to top)


Quickstart

CLI

# Zero-config demo
binex hello

# Run a workflow
binex run examples/simple.yaml

# Inspect the run
binex debug latest
binex trace latest

Web UI

binex ui

Create a Workflow

name: research-pipeline
nodes:
  input:
    agent: "human://input"
    outputs: [output]

  planner:
    agent: "llm://gemini/gemini-2.5-flash"
    system_prompt: "Break this topic into research questions"
    depends_on: [input]
    outputs: [output]

  researcher:
    agent: "llm://openrouter/google/gemma-3-27b-it:free"
    system_prompt: "Investigate and report findings"
    depends_on: [planner]
    outputs: [output]

  output:
    agent: "human://output"
    depends_on: [researcher]
    outputs: [output]

(back to top)


Features

Agent Adapters

Prefix Description
local:// In-process Python callable
llm:// LLM via LiteLLM (40+ providers)
a2a:// Remote agent via A2A protocol
human://input Free-text input from user
human://approve Approval gate with conditional branching
human://output Display results to user
langchain:// LangChain Runnable (plugin)
crewai:// CrewAI Crew (plugin)
autogen:// AutoGen Team (plugin)

CLI Commands

Command Description
binex run Execute a workflow
binex ui Launch Web UI
binex debug Post-mortem inspection
binex trace Execution timeline
binex replay Re-run with agent swaps
binex diff Compare two runs
binex cost show Cost breakdown per node
binex explore Interactive TUI dashboard
binex scaffold Generate workflow from DSL
binex export Export to CSV/JSON
binex doctor System health check
binex hello Zero-config demo

LLM Providers

OpenAI · Anthropic · Google Gemini · Ollama · OpenRouter · Groq · Mistral · DeepSeek · Together AI

Built With

Python React FastAPI TypeScript Tailwind

(back to top)


Examples

Example What it demonstrates
simple.yaml Minimal two-node pipeline
diamond.yaml Diamond dependency pattern
fan-out-fan-in.yaml Parallel execution with aggregation
human-in-the-loop.yaml Approval gates and conditional branching
multi-provider-demo.yaml Multiple LLM providers in one workflow
ollama-research.yaml Full research pipeline with Ollama + OpenRouter
langchain-summarizer.yaml LangChain Runnable in a pipeline
crewai-research-crew.yaml CrewAI Crew as a workflow node
autogen-coding-team.yaml AutoGen Team for code generation

(back to top)


Architecture

src/binex/
├── adapters/        # Agent backends (local, LLM, A2A, human, frameworks)
├── cli/             # Click CLI commands
├── graph/           # DAG construction + topological scheduling
├── models/          # Pydantic v2 domain models
├── plugins/         # Plugin registry for custom adapters
├── prompts/         # 121 built-in prompt templates
├── runtime/         # Orchestrator, dispatcher, replay engine
├── stores/          # SQLite execution + filesystem artifacts
├── trace/           # Debug, lineage, timeline, diffing
├── ui/              # FastAPI backend + React frontend
│   ├── api/         # 20 REST endpoints
│   └── static/      # Pre-built React app
└── workflow_spec/   # YAML loader + validator

(back to top)


Documentation

Full docs at alexli18.github.io/binex

(back to top)


Contributing

Contributions are welcome! If you find this useful:

  • Star the repo — it takes 1 second and helps more than you know
  • Open issues — tell me what's broken or what you need
  • Submit PRs — let's build this together

I'm a solo developer building this in the open. Every star, issue, and PR makes a real difference.

  1. Fork the Project
  2. Create your Feature Branch (git checkout -b feature/AmazingFeature)
  3. Commit your Changes (git commit -m 'Add some AmazingFeature')
  4. Push to the Branch (git push origin feature/AmazingFeature)
  5. Open a Pull Request

(back to top)


License

Distributed under the MIT License. See LICENSE for more information.

(back to top)


Built by a solo dev who believes AI agents shouldn't be black boxes.
No cloud. No telemetry. No surprises. Just debuggable AI workflows.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

binex-0.5.0-py3-none-any.whl (306.8 kB view details)

Uploaded Python 3

File details

Details for the file binex-0.5.0-py3-none-any.whl.

File metadata

  • Download URL: binex-0.5.0-py3-none-any.whl
  • Upload date:
  • Size: 306.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.2

File hashes

Hashes for binex-0.5.0-py3-none-any.whl
Algorithm Hash digest
SHA256 15761d54d26a3f4c99b84f212dd243513fcf0cd4c1a71e63c9d2a7dace42286b
MD5 a1f559b566a23b6d32a3fca7ac8894b3
BLAKE2b-256 ec930651314cf292371a2baa1143c9073932e73039adc6893d0e03f987f9321a

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page