Skip to main content

AI agent that turns natural language into executable automation. 412 batteries included.

Project description

flyto-ai

flyto-ai

Natural language → executable automation workflows

Most AI agents have the LLM write shell commands and pray. flyto-ai uses 412 pre-built, schema-validated modules instead.

PyPI Python License


The Problem

AI agents like open-interpreter and OpenClaw have the LLM generate shell commands or raw code on every run. This means:

  • Non-deterministic — the same prompt can produce different commands each time
  • No validation — wrong flags, hallucinated APIs, subtle bugs only found at runtime
  • Not reusable — each execution is ephemeral, nothing saved for next time
  • Expensive — LLM spends tokens figuring out how to execute, not just what to execute

The Fix

flyto-ai flips the model: the LLM never writes code. It searches and selects from 412 pre-built modules, fills in parameters (validated against schemas), and executes them deterministically. Every run produces a reusable YAML workflow.

❯ scrape the title from example.com

Result: "Example Domain"
name: Scrape Title
params:
  url: "https://example.com"
steps:
  - id: launch
    module: browser.launch
  - id: goto
    module: browser.goto
    params:
      url: "${{params.url}}"
  - id: extract
    module: browser.extract
    params:
      selector: "h1"

Quick Start

pip install flyto-ai
playwright install chromium     # download browser for web automation
export OPENAI_API_KEY=sk-...   # or ANTHROPIC_API_KEY
flyto-ai

One install, one command — interactive chat with 412 automation modules, browser automation, and self-learning blueprints.

flyto-ai demo

How It's Different

The core difference is what the LLM does during execution:

open-interpreter / OpenClaw flyto-ai
LLM's job Write shell/Python code from scratch Select modules + fill params
Execution subprocess.run(llm_output) execute_module("browser.extract", {validated_params})
Validation None — errors at runtime Schema validation before execution
Determinism Same prompt → different code Same module + params → same result
Output One-time result Result + reusable YAML workflow
Learning None Self-learning blueprints (zero LLM replay)
Cost per replay Full LLM inference again $0 (saved blueprint, no LLM)

Benchmark: "Scrape the title from example.com"

open-interpreter flyto-ai
Tokens used ~8K (writes Python + subprocess) ~2K (search → schema → execute)
Execution time ~12s (LLM generates code + runs) ~8s (LLM selects modules + runs)
Second run ~12s (same cost, regenerate code) ~0.5s (blueprint replay, zero LLM)
Reusable output No Yes (YAML workflow)
Deterministic No Yes

Why flyto-ai?

aider open-interpreter flyto-ai
Output Code changes (git diff) One-time code execution Results + reusable YAML workflows
Tools Your codebase Raw Python/JS/Shell 412 pre-built modules
Learns No No Yes — self-learning blueprints
Reusable Yes (code) No (ephemeral) Yes (save, share, schedule)
Webhook/API No No Yes
For Developers Power users Developers & ops automation
License Apache-2.0 AGPL-3.0 Apache-2.0

Use Cases

Web Scraping

❯ extract all product names and prices from example-shop.com/products
name: Scrape Products
params:
  url: "https://example-shop.com/products"
steps:
  - id: launch
    module: browser.launch
  - id: goto
    module: browser.goto
    params:
      url: "${{params.url}}"
  - id: extract
    module: browser.extract
    params:
      selector: ".product"
      fields:
        name: ".product-name"
        price: ".product-price"

Form Automation

❯ log in to staging.example.com, fill the contact form, and take a screenshot
name: Fill Contact Form
steps:
  - id: launch
    module: browser.launch
  - id: login
    module: browser.login
    params:
      url: "https://staging.example.com/login"
      username_selector: "#email"
      password_selector: "#password"
      submit_selector: "button[type=submit]"
  - id: fill
    module: browser.form
    params:
      url: "https://staging.example.com/contact"
      fields:
        name: "Test User"
        message: "Hello from flyto-ai"
  - id: proof
    module: browser.screenshot

API Monitoring + Notification

❯ check if https://api.example.com/health returns 200, if not send a Slack message
name: Health Check Alert
params:
  endpoint: "https://api.example.com/health"
steps:
  - id: check
    module: http.get
    params:
      url: "${{params.endpoint}}"
  - id: notify
    module: notification.slack
    params:
      webhook_url: "${{params.slack_webhook}}"
      message: "Health check failed: ${{steps.check.status_code}}"
    condition: "${{steps.check.status_code}} != 200"

412 Batteries Included

Powered by flyto-core — 412 automation modules across 55 categories:

Category Modules Examples
Browser 39 launch, goto, click, type, extract, screenshot, wait
Atomic 35 reusable building-block operations
Flow 23 conditionals, loops, branching, error handling
Cloud 14 S3, GCS, cloud storage and APIs
Data 13 JSON, CSV, parsing, transformation
Array 12 filter, map, sort, flatten, unique
String 11 split, replace, template, regex, slugify
Productivity 10 email, calendar, document integrations
Image 9 resize, crop, convert, watermark, compress
HTTP / API 9 GET, POST, download, upload, GraphQL
Notification 9 email, Slack, Telegram, webhook
+ 44 more 200+ database, crypto, docker, k8s, testing, ...

Browse available modules:

flyto-ai version   # Shows installed module count

Self-Learning Blueprints

The agent remembers what works. Good workflows are automatically saved as blueprints — reusable patterns that make future tasks faster and free.

First time:  "screenshot example.com" → 15s (discover modules, build from scratch)
Second time: "screenshot another.com" → 3s  (reuse learned blueprint, zero LLM cost)

How it works (closed-loop, no LLM involved):

  1. Execution succeeds with 3+ steps → auto-saved as blueprint (score 70)
  2. Blueprint reused successfully → score +5
  3. Blueprint fails → score -10
  4. Score < 10 → auto-retired, never suggested again
flyto-ai blueprints                             # View learned blueprints
flyto-ai blueprints --export > blueprints.yaml  # Export for sharing

CLI

flyto-ai                                     # Interactive chat — executes tasks directly
flyto-ai chat "scrape example.com"           # One-shot execute mode
flyto-ai chat "scrape example.com" --plan    # YAML-only mode (don't execute)
flyto-ai chat "take screenshot" -p ollama    # Use Ollama (no API key needed)
flyto-ai chat "..." --webhook https://...    # POST result to webhook
flyto-ai serve --port 8080                   # HTTP server for triggers
flyto-ai blueprints                          # List learned blueprints
flyto-ai version                             # Version + dependency status

Interactive Mode

Just run flyto-ai — multi-turn conversation with up/down arrow history:

$ flyto-ai

  _____ _       _        ____       _    ___
 |  ___| |_   _| |_ ___ |___ \     / \  |_ _|
 | |_  | | | | | __/ _ \  __) |   / _ \  | |
 |  _| | | |_| | || (_) |/ __/   / ___ \ | |
 |_|   |_|\__, |\__\___/|_____|  /_/   \_\___|
           |___/

  v0.6.0  Interactive Mode
  Provider: openai  Model: gpt-4o  Tools: 412

  ⏵⏵ execute · openai/gpt-4o · 412 tools
❯ scrape the title from example.com

  ○ browser.launch
  ○ browser.goto
  ○ browser.extract

  The title of example.com is: **Example Domain**

  3 executed · 5 tool calls

  ⏵⏵ execute · openai/gpt-4o · 412 tools · 1 msgs
❯ now also take a screenshot

❯ /mode
Switched to: plan-only (YAML output)

Commands: /clear, /mode, /history, /version, /help, /exit

Webhook & HTTP Server

Send results anywhere:

flyto-ai chat "scrape example.com" --webhook https://hook.site/xxx

Accept triggers from anywhere:

flyto-ai serve --port 8080

# From Slack, n8n, Make, or any HTTP client:
curl -X POST http://localhost:8080/chat \
  -H "Content-Type: application/json" \
  -d '{"message": "take a screenshot of example.com"}'

# Execute mode (default) or plan-only:
curl -X POST http://localhost:8080/chat \
  -H "Content-Type: application/json" \
  -d '{"message": "scrape example.com", "mode": "yaml"}'

Python API

from flyto_ai import Agent, AgentConfig

agent = Agent(config=AgentConfig.from_env())

# Execute mode (default) — runs modules and returns results
result = await agent.chat("extract all links from https://example.com")
print(result.message)            # Result + YAML workflow
print(result.execution_results)  # Module execution results

# Plan-only mode — generates YAML without executing
result = await agent.chat("extract all links from example.com", mode="yaml")
print(result.message)            # YAML workflow only

Multi-Provider

Works with any LLM provider:

export OPENAI_API_KEY=sk-...          # OpenAI models
export ANTHROPIC_API_KEY=sk-ant-...   # Anthropic models
flyto-ai chat "..." -p ollama         # Local models (Llama, Mistral, etc.)
flyto-ai chat "..." --model <name>    # Any specific model

Security

  • Workflows are auditable — YAML is human-readable, reviewable, and version-controllable
  • Module policies — whitelist/denylist categories (e.g. block file.* or database.*)
  • Sensitive param redaction — API keys and passwords are masked in tool call logs
  • Local-first — blueprints stored in local SQLite, nothing sent to third parties
  • Webhook output — structured JSON only, no raw credentials in payload

Architecture

User message
  → LLM (OpenAI / Anthropic / Ollama)
    → Function calling: search_modules, get_module_info, execute_module, ...
      → 412 flyto-core modules (schema-validated, deterministic)
      → Self-learning blueprints (closed-loop, zero LLM)
      → Browser page inspection
    → Execute mode: run modules, return results + YAML
    → Plan mode: YAML validation loop (auto-retry on errors)
  → Structured output (results + reusable workflow)

Environment Variables

Variable Description
FLYTO_AI_PROVIDER openai, anthropic, or ollama
FLYTO_AI_API_KEY API key (or use provider-specific vars below)
FLYTO_AI_MODEL Model name override
OPENAI_API_KEY Fallback for OpenAI provider
ANTHROPIC_API_KEY Fallback for Anthropic provider
FLYTO_AI_BASE_URL Custom API endpoint (OpenAI-compatible)

License

Apache-2.0 — use it commercially, fork it, build on it.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

flyto_ai-0.9.3.tar.gz (181.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

flyto_ai-0.9.3-py3-none-any.whl (105.7 kB view details)

Uploaded Python 3

File details

Details for the file flyto_ai-0.9.3.tar.gz.

File metadata

  • Download URL: flyto_ai-0.9.3.tar.gz
  • Upload date:
  • Size: 181.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.10.6

File hashes

Hashes for flyto_ai-0.9.3.tar.gz
Algorithm Hash digest
SHA256 0f1dee6c9650417fb40d73f21c6ef47d70658172e93cc7dd5a524e17d3d828f7
MD5 f7891bf11cfea6157dd8da0667f52ec1
BLAKE2b-256 3042b6a423f1038046d684e2fac51957c85d0e50021cf78308c17d3781a750a9

See more details on using hashes here.

File details

Details for the file flyto_ai-0.9.3-py3-none-any.whl.

File metadata

  • Download URL: flyto_ai-0.9.3-py3-none-any.whl
  • Upload date:
  • Size: 105.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.10.6

File hashes

Hashes for flyto_ai-0.9.3-py3-none-any.whl
Algorithm Hash digest
SHA256 d0a93f83c3941c1188be1888b07d22b9b56ca27d83288a4666cd8267c5acfb5c
MD5 12be17cdc94414fb847ffadda726c748
BLAKE2b-256 320b16cee6a006d399869ad798c469dc4d8aef2ce00ed4a993fc0456a327b004

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page