Synapse - A Multi Agent Orchestrator
Project description
Synapse AI — Multi-Agent Orchestration Platform
Build AI workflows that actually ship.
Wire agents, tools, and LLMs into deterministic pipelines — without the framework lock-in. Synapse is an open-source platform for creating, connecting, and orchestrating AI agents powered by any LLM — local or cloud. Agents use real tools: browsing the web, querying databases, executing code, reading files, managing emails, trading stocks, and anything else you can expose through an MCP server, a webhook, or a Python script — if you can write it, agents can use it.
Prerequisites
- Python 3.11+
- Node.js 22+
- uvx —
pip install uv(used to run MCP servers)
Don't have these? The setup script will attempt to install any missing prerequisites automatically.
Install
Quick Setup Script (recommended)
The easiest way to get started. Clones the repository, installs all dependencies, and starts both servers automatically.
macOS / Linux:
curl -sSL https://raw.githubusercontent.com/naveenraj-17/synapse-ai/main/setup.sh | bash
Windows (PowerShell):
irm https://raw.githubusercontent.com/naveenraj-17/synapse-ai/main/setup.ps1 | iex
npm
npm install -g synapse-orch-ai
synapse
pip
pip install synapse-orch-ai
synapse
CLI
Once installed, use the synapse command to manage the server:
synapse start # start backend + frontend, open browser
synapse stop # stop background processes
synapse upgrade # upgrade to the latest version
synapse uninstall # remove Synapse, wipe ~/.synapse, and uninstall the package
What Makes Synapse Different
Most agent frameworks hand you a loop, a few toy tools, and a tutorial. Synapse is a production-grade orchestration platform built for the real thing:
Cut Costs Without Cutting Quality
Run a different model at every step. Use a fast, cheap model for routing and classification; switch to a powerful model only for the steps that need it. One orchestration, many models — you control exactly where the compute goes.
Workflows That Actually Do What You Designed
Orchestrations are strict DAGs. Execution follows the exact path you defined — no surprises, no hallucinated detours. For steps where the next action is already known (fetch this, parse that, write here), use Tool and LLM steps instead of full agents: zero reasoning overhead, deterministic output, and far cheaper to run.
Turn Anything Into a Tool
- Any Python program → drop it in, it becomes a sandboxed agent tool
- Any API or webhook → describe its schema, agents call it natively
- Any MCP server → local subprocess or remote HTTP, connected in seconds
- Any orchestration → promote it to an agent; chain orchestrations like functions
Never Blocked on a Human Decision
Human steps pause execution mid-workflow and wait. When the person responds — via the UI, Slack, Telegram, or any connected messaging channel — the run resumes exactly where it left off. No polling, no timeouts you didn't set.
Run It Anywhere, Own Your Data
Full local operation with Ollama. Or mix: local models for some agents, cloud APIs for others. No vendor lock-in on models, no data sent anywhere you didn't choose. Persistent vault stores files across agent sessions on your machine.
Built-In Scheduling & Messaging
Schedule any agent or orchestration to run on a cron or interval. Results are pushed directly to Slack, Discord, Telegram, Teams, or WhatsApp — with multi-agent mode so users can switch agents mid-chat.
Under the Hood
| Capability | Detail |
|---|---|
| Multi-model orchestrations | Per-step model override — mix Gemini Flash, Claude Opus, and local Ollama in one workflow |
| Orchestrations as agents | Promote any orchestration to an agent; nest pipelines inside pipelines |
| Deterministic Tool steps | Skip the ReAct loop entirely — call a specific tool directly with state values |
| Resumable human gates | Human steps survive server restarts; runs pick up exactly where they paused |
| Docker-sandboxed Python | Agents write and execute Python in an isolated container — safe by default |
| Stealth web scraping | Built-in anti-bot evasion; works on LinkedIn, financial sites, JS-heavy pages |
| Semantic code search | Index any repo; agents query it by natural language and get file + line results |
| Cost limits per run | Set a max-spend per orchestration run — execution halts if the budget is hit |
| 5 messaging platforms | Slack, Discord, Telegram, Teams, WhatsApp — with per-channel agent binding |
| 14+ LLM providers | Cloud, local, and CLI providers; no API key needed for Claude/Gemini/Codex CLI |
| Import/Export packs | Portable bundles of agents + orchestrations + MCP configs; 3 curated starter packs |
| AI Builder | Chat with a meta-agent that designs and materializes orchestrations for you |
Synapse UI
https://github.com/user-attachments/assets/7a5ab42c-5fae-4f13-876c-13aa9b5a0366
Synapse Orchestration Demo
Content Writing Orchestration
This demo showcases a multi-agent content writing orchestration pipeline. The agents autonomously open a browser, research a user-provided topic, draft the content in a Google Doc, and return the shared link. By default, the worker agents utilize gemini-3-flash-preview, while the evaluator agents use gemini-3.1-pro-preview. Each agent can be configured with different models based on your specific requirements. (Note: The video is sped up 2x to fit on GitHub.)
https://github.com/user-attachments/assets/4eec5db8-70d0-47b6-8608-f52b1f7b7d68
Autonomous Code Development & PR Creation
This demo highlights a multi-agent software development system that writes code and generates pull requests on its own. A human-in-the-loop step is integrated into each stage, allowing you to review and confirm the agents' actions before the system finalizes the PR and outputs the repository link.
https://github.com/user-attachments/assets/95a511e1-e3e9-4812-b9ca-f7f4c28ef80f
Native Orchestration Builder
Instead of manually dragging and dropping to create your flow, you can now just chat with the builder. Tell it what kind of orchestration you want, and the AI will build the DAG for you. Once it maps it out, you can just start running it immediately.
https://github.com/user-attachments/assets/282cc99d-cdea-4ad0-b648-f22112c6e295
The Tool Ecosystem
Synapse agents are powerful because of what they can do. Every tool is a separate MCP process — isolated, composable, and safe.
Native Tool Servers
These run automatically when Synapse starts:
| Tool Server | What It Does |
|---|---|
| Sandbox | Execute Python code in an isolated Docker container (512 MB RAM, 1 CPU). Pre-loaded with pandas, numpy, matplotlib, scikit-learn, requests, and more. Read/write files in the persistent vault. |
| Vault | Persistent file storage for agents. Create, read, update, patch, and list files across sessions. JSON deep-merge, text find-replace, and directory listing built in. |
| SQL Agent | Connect to any database (PostgreSQL, MySQL, SQLite). List tables, introspect schemas, run read queries. Supports any SQLAlchemy-compatible connection string. |
| Browser | Full browser automation via Playwright MCP. Navigate pages, click, fill forms, take screenshots, extract content. Powered by Chromium. |
| PDF Parser | Extract text and tables from any PDF by URL. Tables converted to Markdown. Page-by-page extraction. |
| Excel Parser | Parse .xlsx files from URL. Multi-sheet support. Converts all sheets to Markdown tables. |
| Collect Data | Generate dynamic forms that pause execution and collect user input. Supports text, number, email, date, phone, and option fields. |
| Time | Natural language date/time parsing. Handles relative offsets, weekday targets, timezone conversions, and complex expressions like "next Friday at 3pm EST". |
| Code Search | Semantic code search across indexed repositories using vector embeddings. Search by natural language query, get back relevant code snippets with file paths and line numbers. |
| Web Scraper | Powerful web scraping powered by crawl4ai. Scrape any URL to clean markdown, extract structured data with CSS schemas, crawl multiple URLs in parallel, capture screenshots, handle infinite-scroll pages, and run multi-step authenticated sessions. Built-in stealth mode bypasses anti-bot protections — works on LinkedIn, financial sites, and JavaScript-heavy pages. |
Built-in MCP Servers
Enabled automatically when configured:
| Server | What It Does |
|---|---|
Filesystem (@modelcontextprotocol/server-filesystem) |
Full read/write access to your local code repositories. Configure which paths to expose in Settings → Repos. |
Google Workspace (workspace-mcp) |
Gmail (read, search, send), Google Drive (list, read, create files), and Google Calendar (events, scheduling). One-click OAuth setup in Settings. |
Playwright (@playwright/mcp) |
Browser control — already included in the native Browser tool above, available separately for headless automation. |
Sequential Thinking (npx @modelcontextprotocol/server-sequential-thinking) |
Structured step-by-step reasoning for complex, multi-stage problems. Agents break tasks into explicit thought chains before acting. Enabled by default. |
Memory (npx @modelcontextprotocol/server-memory) |
Persistent knowledge graph memory across sessions. Agents store and retrieve facts, relationships, and context between runs. Enabled by default. |
Remote MCP Servers
Connect to any MCP server over the network — no code needed. Synapse supports native OAuth and Personal Access Token (PAT) authentication.
To add a remote server:
- Open Settings → MCP Servers
- Click the Remote (URL) tab at the top of the form
- Optionally select a preset (Vercel, GitHub Copilot, Jira, Zapier, Figma, Fetch) to auto-fill the URL and token fields
- Enter a Server Name and the Server URL
- Bearer Token / PAT — leave empty to use OAuth (a browser window will open for authorization), or paste a personal access token for PAT-based servers (GitHub, Figma)
- Click Connect Server
Synapse prefixes external tools with <server-unique-name>__ followed by the tool name to prevent naming collisions. Any MCP-compatible API becomes an agent tool instantly.
Find more on the MCP servers registry.
Local (stdio) MCP Servers
For servers that run as local processes, click the Local (stdio) tab and enter the command and arguments:
Command: uvx
Arguments: mcp-server-git
Use the Git preset to auto-fill this. Add environment variables (API keys, secrets) directly in the form — no config file editing required.
Custom Tools (n8n & Python)
Turn any automation workflow or Python logic into an agent tool.
- Build a workflow in n8n (or any webhook-compatible tool), or expose an API via a Python program
- Add it to Synapse in Settings → Custom Tools
- Your agent now has that tool — it sees the name, description, and schema, and calls it like any other tool
This is the fastest way to give agents access to internal APIs, proprietary systems, or multi-step processes that you've already automated. n8n's 400+ node library or your own custom Python logic becomes your agent's extended toolkit.
Building Agents
Create specialized agents in Settings → Agents. Each agent is an independent ReAct loop with its own:
- System prompt — define its persona, expertise, and constraints
- Tool selection — give it access to all tools, or restrict to a specific subset
- Model override — run different agents on different models (e.g., fast model for routing, capable model for analysis)
- Code repositories — link repos for semantic code search and filesystem access
- LLM provider — mix local Ollama models with cloud APIs per agent
Example: Research Agent
{
"name": "Research Agent",
"description": "Deep research using web browsing and document parsing",
"tools": ["browser_navigate", "browser_snapshot", "parse_pdf", "parse_xlsx", "vault_write"],
"system_prompt": "You are a thorough research analyst. For any research task: browse primary sources, extract key data, parse any documents you find, and save a structured report to the vault."
}
Example: Data Agent
{
"name": "Data Agent",
"description": "Analyzes data files and databases, produces reports",
"tools": ["list_tables", "get_table_schema", "run_sql_query", "execute_python", "vault_write", "vault_read"],
"system_prompt": "You are a data analyst. Explore the database schema, write SQL queries to extract insights, then use Python (pandas/matplotlib) to analyze and visualize results. Save all outputs to the vault."
}
Example: Developer Agent
{
"name": "Strict Developer",
"description": "Writes production-ready code, creates APIs, and runs self-correcting tests",
"tools": ["execute_python", "mcp_github", "mcp_slack", "vault_write", "vault_read"],
"system_prompt": "You are a senior backend engineer. Write robust, functional code, execute it using the Python tool to verify logic, and save the final output to the vault."
}
AI Builder
The Synapse AI Builder is a native multi-agent orchestration that helps you design, review, and create new orchestrations through a guided conversation — no canvas required.
Describe what you want to build in plain language. The builder will:
- Understand your requirements (and ask clarifying questions if needed)
- Draft a plan — structured markdown with a step-by-step breakdown and an ASCII flow diagram
- Present the plan for your review — approve it or request revisions in plain text
- Create any new sub-agents the orchestration needs (if enabled)
- Materialise the orchestration by calling
create_orchestrationorupdate_orchestration - Confirm with a friendly summary and the new orchestration ID
The Builder can also edit existing orchestrations — point it at an orchestration you are viewing and describe your changes.
The builder is seeded automatically at startup and is always available in the orchestration picker.
Orchestrating Agents
Individual agents are powerful. Orchestrations are transformative.
An orchestration is a directed graph (DAG) of steps — you wire agents together, add routing logic, run things in parallel, loop over datasets, and checkpoint for human review. Build them visually on the canvas or define them in JSON.
Step Types
| Step | What It Does |
|---|---|
| Agent | Run an agent's full ReAct loop. Pass context from shared state as input. Capture the result as an output key. |
| LLM | Make a direct LLM call without spinning up a full agent loop. Use for single-shot generation, summarization, classification, or prompt templating against shared state. Faster and cheaper than a full agent step when tool use isn't needed. |
| Tool | Execute a specific MCP tool directly — no agent reasoning, no loop. Pass inputs from shared state, write the raw tool output back to state. Ideal for deterministic data-fetching steps (e.g., run a SQL query, read a vault file, call an API). |
| Evaluator | Ask an LLM to make a routing decision. Maps decision labels to next steps. Use this to branch based on analysis results. |
| Parallel | Run multiple agent branches. Each branch runs sequentially (respects shared resources like browser). |
| Merge | Combine outputs from parallel branches. Strategies: list (accumulate), concat (join text), dict (merge objects). |
| Loop | Repeat a set of steps N times. Use with transforms to iterate over lists or refine outputs. |
| Transform | Execute arbitrary Python against the shared state dict. Reshape data, compute values, filter lists. |
| Human | Pause and ask a human for input via a generated form. Execution resumes when the user responds. Fully resumable. |
| Extract JSON | Parse JSON out of any text — handles raw JSON, markdown code fences, and multiple objects (stored as an array). No LLM call. Perfect for pulling structured data out of an agent's raw output. |
Render a text or Markdown template with {state.key} interpolation and store it in the shared state. Use for building formatted summaries, reports, or notification bodies without an LLM call. |
|
| IF / Else | Evaluate a Python expression against the shared state and branch to one of two steps — true path or false path. Supports dot-notation (state.result.flag). Missing keys evaluate to None. No LLM call. |
| Switch | Match a Python expression's string result against a set of named cases. Each case routes to a different step; unmatched values fall through to the default route. No LLM call. |
| End | Finalize the workflow. |
Deterministic Control-Flow Steps
Four step types execute without any LLM call — they are fast, free, and completely predictable. Use them to add control flow and data handling between your agent steps.
Extract JSON
Finds and parses JSON from raw text. Works with:
- Plain JSON objects / arrays
- Markdown code fences (
```json ... ```) - Multiple JSON blocks in a single string (stored as an array)
Input key: llm_raw_output (e.g. "The answer is: ```json\n{\"score\": 8}\n```")
Output key: parsed (→ { "score": 8 })
Renders a Markdown or plain-text template with {state.key} and {state.key.nested} placeholders resolved from the shared state, then stores the result.
print_content: "# Report\n\nScore: {state.score}\nCategory: {state.category}"
output_key: report_text
IF / Else
Evaluates a Python expression against the shared state and branches to one of two steps. Dot-notation is supported — missing keys are None.
if_condition: state.score > 7
if_true_step_id: step_approve
if_false_step_id: step_reject
Safe built-ins only (len, str, int, float, bool, list, dict, max, min, abs, round, any, all). No imports.
Switch
Converts a Python expression to a string and matches it against named cases. Unmatched values fall through to switch_default_step_id.
switch_expression: state.category
switch_cases:
"sports" → step_sports_handler
"politics" → step_politics_handler
"science" → step_science_handler
switch_default_step_id: step_general_handler
Tip: Chain these steps to build lightweight classification pipelines — use an LLM step to classify, an Extract JSON step to parse its output, and a Switch step to route — all without extra LLM calls.
Shared State
Every step reads from and writes to a shared state dictionary. Define the schema upfront:
"state_schema": {
"query": { "type": "string", "description": "Initial user query" },
"research_results": { "type": "string", "description": "Raw research output" },
"analysis": { "type": "string", "description": "Structured analysis" },
"approved": { "type": "boolean", "default": false }
}
Steps use input_keys to pull from state and output_key to write back. This is how agents hand off work to each other.
Example: End-to-End Research → Report Orchestration
Here's a complete orchestration that combines 5 agents to go from a question to a published report with human approval:
User Query
│
▼
[1. Research Agent] → Browses web, parses PDFs, saves raw findings to vault
│ output: research_raw
▼
[2. Parallel Step]
├── [3. Data Agent] → Pulls supporting data from SQL, runs Python analysis
└── [4. Fact Checker] → Cross-references key claims via browser
│ output: data_analysis, verified_facts
▼
[5. Merge] → Combines data_analysis + verified_facts
│
▼
[6. Writer Agent] → Synthesizes all inputs into structured report, saves to vault
│ output: report_draft
▼
[7. Quality Evaluator] → Routes: "approved" → Human Review | "needs_revision" → Writer Agent
│
▼
[8. Human Review] → Shows draft, collects approval or revision notes
│
▼
[9. Publisher Agent] → Sends report via email (Gmail MCP), posts to Drive
│
▼
[END]
This orchestration:
- Runs 3 agents in parallel (saves time)
- Routes automatically based on quality assessment
- Loops the writer if revisions are needed
- Pauses for human approval before publishing
- Uses vault to pass files between agents
- Publishes via Gmail and Google Drive
Build this visually on the canvas in about 10 minutes.
Example: Stock Analysis Orchestration
The included "Stock Intraday Trading" orchestration shows how to combine market data, risk analysis, and human decisions:
[1. Portfolio Analyzer] → Checks current positions via Zerodha MCP
│
▼
[2. Login Router] → Evaluator: logged in? → continue | not logged in? → prompt user
│
▼
[3. Parallel Analysis]
├── [NSE Stock Analyzer] → Technical analysis on watchlist
├── [Beta Data Fetcher] → Fetches beta/volatility data
└── [Current Events Agent] → Browses news, checks sentiment
│
▼
[4. Merge + Strategy Transform] → Python transform: compute risk-adjusted scores
│
▼
[5. Human Approval] → Shows recommended trades, waits for confirmation
│
▼
[END]
Schedules
Automate agent and orchestration runs on a recurring schedule.
- Interval — run every N minutes/hours/days (e.g. monitor a feed, poll an API)
- Cron / Fixed Time — run at specific times (e.g. every day at 9 AM for a morning standup report)
- Prompt — The prompt is what the agent will receive each time the schedule fires. Think of it as a standing instruction.
- Messaging notifications — If the selected agent has a connected messaging channel (Slack, Telegram, etc.), the result is sent there automatically after each run.
Messaging
Connect your agents to Telegram, Discord, Slack, Teams, or WhatsApp.
Enable Multi-Agent Mode in a channel so users can switch agents mid-chat using /agent <name> and list them with /agents. The channel's bound agent is the default.
Import / Export
Export your orchestrations, agents, MCP servers, and tools as a portable bundle, or import one from another Synapse instance.
Example Packs: Synapse includes curated collections of agents, orchestrations, and MCP servers. Select a pack to preview what will be imported before committing:
- Starter Pack — Get up and running fast, includes a Personal Assistant with full tool access and a Web Research Agent.
- Developer Pack — Built for engineering teams, includes a Code Review Agent, Software Engineer Agent, QA Engineer, and a Dev base orchestration.
- Productivity Pack — Business and content power-users, includes a Data Analyst, Content Writer, Jira Analyst, Slack Notifier.
Vault Management
The Vault is a persistent file directory (data/vault/) that acts as shared storage for all your agents.
- UI File Explorer — Manage your vault directly in Settings → Vault. Create, edit, and delete
.md,.json, and.txtfiles in a full-featured markdown and JSON editor. - Context Injection — Instantly reference files inside agent system prompts or orchestration templates by typing
@. The UI provides an intelligent dropdown that searches the vault, allowing you to seamlessly embed documents, guides, or skill configurations into LLM contexts as@[path/to/file.md]. - Agent Access — Agents have built-in tool access to dynamically read, write, and patch files in the vault across sessions.
Configuration
Supported LLM Providers
| Provider | Mode | Model prefix | Notes |
|---|---|---|---|
| Ollama | Local | (none — bare model name) | Any model pulled via ollama pull. Default: mistral-nemo |
| Anthropic | Cloud | claude- |
Claude 3.5, Claude 3 Opus, Claude 3.7 Sonnet, etc. |
| OpenAI | Cloud | gpt- |
GPT-4o, GPT-4 Turbo, o1, o3-mini, etc. |
| Gemini | Cloud | gemini- / gemma- |
Gemini 1.5 Pro, Gemini 2.0 Flash, etc. |
| xAI (Grok) | Cloud | grok- |
Grok-2, Grok-3, Grok-3 Mini. Set XAI_API_KEY. |
| DeepSeek | Cloud | deepseek- |
DeepSeek-V3, DeepSeek-R1 (reasoning). Set DEEPSEEK_API_KEY. |
| AWS Bedrock | Cloud | bedrock. |
Any Bedrock model (Converse API). Set AWS credentials or a Bedrock API key in Settings. |
| Ollama v1 Compatible | Cloud | oaic.<model> |
Any cloud OpenAI-compatible endpoint (OpenRouter, Together AI, Fireworks, etc.). Configure Base URL + API key in Settings → Model. |
| Local v1 Compatible | Local | locv1.<model> |
Any local OpenAI-compatible server (vLLM, LM Studio, Jan, Ollama /v1, etc.). Configure Base URL (and optional key) in Settings → Model. |
| Claude CLI | CLI | cli.claude |
Requires the Claude Code CLI (claude) installed and authenticated. No API key needed — uses your existing Claude subscription. |
| Gemini CLI | CLI | cli.gemini |
Requires the Gemini CLI (gemini) installed and authenticated. Supports pro and flash variants. |
| Codex CLI | CLI | cli.codex |
Requires the Codex CLI (codex) installed and authenticated. No API key needed — uses your existing OpenAI subscription. |
| GitHub Copilot CLI | CLI | cli.copilot |
Requires the GitHub Copilot CLI extension (copilot) installed and authenticated. No API key needed — uses your GitHub Copilot subscription. |
Switch providers per-agent or globally in Settings → Model.
Star History
Environment Variables
# Copy and edit
cp .env.example .env
| Variable | Default | Description |
|---|---|---|
SYNAPSE_DATA_DIR |
~/.synapse/data |
Where agents store files, memory, and state |
OLLAMA_BASE_URL |
http://127.0.0.1:11434 |
Local Ollama endpoint |
SYNAPSE_BACKEND_PORT |
8765 |
Backend API port |
SYNAPSE_FRONTEND_PORT |
3000 |
Frontend UI port |
BACKEND_URL |
http://127.0.0.1:8765 |
Backend URL as seen by Next.js server (set in Docker) |
CORS_ORIGINS |
http://localhost:3000 |
Allowed CORS origins |
Manual Setup
Backend
cd backend
python3.11 -m venv venv
source venv/bin/activate
pip install -r requirements.txt
python3.11 main.py
Frontend
cd frontend
npm install
npm run dev
Open http://localhost:<SYNAPSE_FRONTEND_PORT> (default: http://localhost:3000)
Upcoming Features (Roadmap)
We are constantly improving Synapse AI. Here are a few features currently in the pipeline:
- Spawn Sub-Agent Tool: Allow agents to natively spawn and delegate tasks to temporary sub-agents mid-execution.
- Compact Conversations: A conversation option optimized to handle large contexts smoothly, compressing message history automatically.
- Global Variable: Support for defining global variables that can be dynamically injected into agent prompts, orchestrations, custom tools, and MCP server environments.
Contributing
See CONTRIBUTING.md for dev setup, architecture details, how to add MCP tool servers, and the PR checklist.
License
Synapse AI is licensed under AGPL v3 to ensure it remains open and free, and to prevent cloud monopolies from offering it as a managed service without contributing back to the community.
AGPL-3.0-only — see LICENSE
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file synapse_orch_ai-1.3.3.tar.gz.
File metadata
- Download URL: synapse_orch_ai-1.3.3.tar.gz
- Upload date:
- Size: 686.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.11.15
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
94695ae87c51da4bb99b714ec67fd8996c6605129e01d346ea3e38735b00657d
|
|
| MD5 |
954fd5bd800664b5d6cc866666a89188
|
|
| BLAKE2b-256 |
3148e60f1dc9b837a1c1fa3868662763425f5b1c1bd901ca679a0ccc4ded8088
|
File details
Details for the file synapse_orch_ai-1.3.3-py3-none-any.whl.
File metadata
- Download URL: synapse_orch_ai-1.3.3-py3-none-any.whl
- Upload date:
- Size: 21.1 MB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.11.15
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
af2229eb64c9fa9f6de75beec5cc090495963f829ccbc084e34f7ec2f3cb4045
|
|
| MD5 |
a8aa845045b65c53b0d050ba72cc6e56
|
|
| BLAKE2b-256 |
51fd69f4706f49cc00b4c768cbeddeb9ebc2dcc3d06a6f02a0af2f23fc9017aa
|