Skip to main content

Sandboxed Claude agents on demand. Run Claude Agent SDK in isolated E2B sandboxes and stream results via SSE.

Project description

Sandstorm

Run AI agents in secure cloud sandboxes. One API call. Zero infrastructure.

Claude Agent SDK E2B OpenRouter Python 3.11+ License: MIT

Hundreds of AI agents running in parallel. Hours-long tasks. Tool use, file access, structured output — each in its own secure sandbox. Sounds hard. It's not.

curl -N -X POST https://your-sandstorm-host/query \
  -d '{"prompt": "Find the top 10 trending Python repos on GitHub today, clone each one, count lines of code, and return a ranked summary as JSON"}'

That's the entire integration. Sandstorm wraps the Claude Agent SDK in isolated E2B cloud sandboxes — the agent installs packages, fetches live data, generates files, and streams every step back via SSE. When it's done, the sandbox is destroyed. Nothing persists. Nothing escapes.

  • Any model via OpenRouter -- swap in DeepSeek R1, Qwen 3, Kimi K2, or any of 300+ models through OpenRouter
  • Full agent power -- Bash, Read, Write, Edit, Glob, Grep, WebSearch, WebFetch -- all enabled by default
  • Safe by design -- every request gets a fresh VM that's destroyed after, with zero state leakage
  • Real-time streaming -- watch the agent work step-by-step via SSE, not just the final answer
  • Configure once, query forever -- drop a sandstorm.json for structured output, subagents, MCP servers, and system prompts
  • File uploads -- send code, data, or configs for the agent to work with

Get Started

# Install from PyPI
pip install duvo-sandstorm

# Or from source
git clone https://github.com/tomascupr/sandstorm.git && cd sandstorm
uv sync
cp .env.example .env   # add ANTHROPIC_API_KEY + E2B_API_KEY
duvo-sandstorm serve    # or: uv run python -m uvicorn sandstorm.main:app --reload

If Sandstorm is useful, consider giving it a star — it helps others find it.

Deploy with Vercel

Table of Contents

Quickstart

Prerequisites

Install

# From PyPI
pip install duvo-sandstorm

# Or from source
git clone https://github.com/tomascupr/sandstorm.git
cd sandstorm
uv sync

Setup

# Set your API keys
cp .env.example .env   # then edit with your keys

# Start the server
duvo-sandstorm serve
# Or with uvicorn directly
uv run python -m uvicorn sandstorm.main:app --reload

# Run your first agent
curl -N -X POST http://localhost:8000/query \
  -H "Content-Type: application/json" \
  -d '{"prompt": "Find the top 10 trending Python repos on GitHub and summarize each in one sentence"}'

E2B Sandbox Template

Sandstorm ships with a public pre-built template (work-43ca/sandstorm) that's used automatically — no build step needed. The template includes Node.js 24, @anthropic-ai/claude-agent-sdk, Python 3, git, ripgrep, and curl.

To customize the template (e.g. add system packages or pre-install other dependencies), edit build_template.py and rebuild:

uv run python build_template.py

CLI

After installing, the duvo-sandstorm (or ds) command is available:

Run an agent

The query command is the default — just pass a prompt directly:

ds "Create hello.py and run it"
ds "Analyze this repo" --model opus
ds "Build a chart" --max-turns 30 --timeout 600
ds "Fetch data" --json-output | jq '.type'

The explicit query subcommand also works: ds query "Create hello.py".

Start the server

ds serve                    # default: 0.0.0.0:8000
ds serve --port 3000        # custom port
ds serve --reload           # auto-reload for development

API keys

Keys are resolved in order: CLI flags > environment variables > .env file in current directory.

# Environment variables (most common)
export ANTHROPIC_API_KEY=sk-ant-...
export E2B_API_KEY=e2b_...

# Or CLI flags
ds "hello" --anthropic-api-key sk-ant-... --e2b-api-key e2b_...

How It Works

Client --POST /query--> FastAPI --> E2B Sandbox (isolated VM)
  <---- SSE stream <---- stdout <-- runner.mjs --> query() from Agent SDK
                                     |-- Bash, Read, Write, Edit
                                     |-- Glob, Grep, WebSearch, WebFetch
                                     '-- subagents, MCP servers, structured output
  1. Your app sends a prompt to POST /query
  2. Sandstorm creates a fresh E2B sandbox with the Claude Agent SDK pre-installed
  3. The agent runs your prompt with full tool access inside the sandbox
  4. Every agent message (thoughts, tool calls, results) streams back as SSE events
  5. The sandbox is destroyed when done -- nothing persists

Features

Structured Output

Configure in sandstorm.json to get validated JSON instead of free-form text:

{
  "output_format": {
    "type": "json_schema",
    "schema": {
      "type": "object",
      "properties": {
        "companies": {
          "type": "array",
          "items": {
            "type": "object",
            "properties": {
              "name": { "type": "string" },
              "funding_total": { "type": "number" },
              "sector": { "type": "string" },
              "url": { "type": "string" }
            },
            "required": ["name", "funding_total", "sector"]
          }
        },
        "files_created": {
          "type": "array",
          "items": { "type": "string" }
        }
      },
      "required": ["companies", "files_created"]
    }
  }
}

The agent works normally (scrapes data, installs packages, writes files), then returns validated JSON in result.structured_output.

Subagents

Define specialized agents in sandstorm.json that the main agent can delegate to:

{
  "agents": {
    "scraper": {
      "description": "Crawls websites and saves structured data to disk.",
      "prompt": "Scrape the target, extract data, and save as JSON to /home/user/output/.",
      "tools": ["Bash", "WebFetch", "Write", "Read"],
      "model": "sonnet"
    },
    "report-writer": {
      "description": "Reads collected data and produces formatted reports.",
      "prompt": "Read all data files, synthesize findings, and generate a PDF report with charts.",
      "tools": ["Bash", "Read", "Write", "Glob"]
    }
  }
}

The main agent spawns subagents via the Task tool when it decides they're needed.

File Uploads

Send files in the request for the agent to work with:

curl -N -X POST https://your-sandstorm-host/query \
  -d '{
    "prompt": "Parse these server logs, find error spikes, and write an incident report",
    "files": {
      "logs/app.log": "2024-01-15T10:23:01Z ERROR [auth] connection pool exhausted\n...",
      "logs/deploys.json": "[{\"sha\": \"a1b2c3\", \"ts\": \"2024-01-15T10:20:00Z\"}]"
    }
  }'

Files are written to /home/user/{path} in the sandbox before the agent starts.

MCP Servers

Attach external tools via MCP in sandstorm.json:

{
  "mcp_servers": {
    "sqlite": {
      "command": "npx",
      "args": ["-y", "@modelcontextprotocol/server-sqlite", "/home/user/data.db"]
    },
    "remote-api": {
      "type": "sse",
      "url": "https://api.example.com/mcp/sse",
      "headers": { "Authorization": "Bearer your-token" }
    }
  }
}
Field Type Description
type string "stdio", "http", or "sse"
command string Command for stdio servers
args string[] Command arguments
url string URL for HTTP/SSE servers
headers object Auth headers for remote servers
env object Environment variables

OpenRouter

Sandstorm works with any model available on OpenRouter -- not just Claude. Run agents powered by GPT-4o, Qwen, Llama, DeepSeek, Gemini, Mistral, or any of 300+ models, all through the same API.

Setup

Add three env vars to .env:

ANTHROPIC_BASE_URL=https://openrouter.ai/api
OPENROUTER_API_KEY=sk-or-...
ANTHROPIC_DEFAULT_SONNET_MODEL=anthropic/claude-sonnet-4  # or any OpenRouter model ID

That's it. The agent now routes through OpenRouter. Your existing ANTHROPIC_API_KEY can stay in .env -- Sandstorm automatically clears it in the sandbox when OpenRouter is active.

Using Open-Source Models

Remap the SDK's model aliases to any OpenRouter model:

# Route "sonnet" to Qwen
ANTHROPIC_DEFAULT_SONNET_MODEL=qwen/qwen3-max-thinking

# Route "opus" to DeepSeek
ANTHROPIC_DEFAULT_OPUS_MODEL=deepseek/deepseek-r1

# Route "haiku" to a fast, cheap model
ANTHROPIC_DEFAULT_HAIKU_MODEL=qwen/qwen3-30b-a3b

Then use the alias in your request or sandstorm.json:

curl -N -X POST http://localhost:8000/query \
  -H "Content-Type: application/json" \
  -d '{"prompt": "Analyze this CSV and build a chart", "model": "sonnet"}'

The agent runs on Qwen, DeepSeek, or whatever you mapped -- with full tool use, file access, and streaming.

Per-Request Keys

Pass openrouter_api_key in the request body for multi-tenant setups:

curl -N -X POST http://localhost:8000/query \
  -d '{"prompt": "...", "openrouter_api_key": "sk-or-...", "model": "sonnet"}'

How It Works

The Claude Agent SDK supports custom API endpoints via ANTHROPIC_BASE_URL. OpenRouter exposes an Anthropic-compatible API, so the SDK sends requests to OpenRouter instead of Anthropic directly. OpenRouter then routes to whatever model you've configured. The ANTHROPIC_DEFAULT_*_MODEL env vars tell the SDK which model ID to send when you use aliases like sonnet or opus.

Compatibility

Most models on OpenRouter support the core agent capabilities (tool use, streaming, multi-turn). Models with strong tool-use support (Claude, GPT-4o, Qwen, DeepSeek) work best. Smaller or older models may struggle with complex tool chains.

Browse available models at openrouter.ai/models.

Configuration

Sandstorm uses a two-layer config system:

Layer What it controls How to set
sandstorm.json Agent behavior -- system prompt, structured output, subagents, MCP servers Config file in project root
API request Per-call -- prompt, model, files, timeout JSON body on POST /query

sandstorm.json

Drop a sandstorm.json in your project root to configure the agent's behavior:

{
  "system_prompt": "You are a due diligence analyst. Write reports to /home/user/output/.",
  "model": "sonnet",
  "max_turns": 20
}

See Structured Output, Subagents, and MCP Servers for advanced configuration.

Field Type Description
system_prompt string Custom instructions for the agent
model string Default model ("sonnet", "opus", "haiku", or full ID)
max_turns integer Maximum conversation turns
output_format object JSON schema for structured output
agents object Subagent definitions
mcp_servers object MCP server configurations

API Keys

Keys can live in .env (set once) or be passed per-request (multi-tenant). Request body overrides .env.

# .env -- set once, forget about it
ANTHROPIC_API_KEY=sk-ant-...
E2B_API_KEY=e2b_...

# Then just send prompts:
curl -N -X POST https://your-sandstorm-host/query \
  -d '{"prompt": "Crawl docs.stripe.com/api and generate an OpenAPI spec as YAML"}'

# Or override per-request:
curl -N -X POST https://your-sandstorm-host/query \
  -d '{"prompt": "...", "anthropic_api_key": "sk-ant-other", "e2b_api_key": "e2b_other"}'

Providers

Sandstorm supports Anthropic (default), Google Vertex AI, Amazon Bedrock, Microsoft Azure, OpenRouter, and custom API proxies. Add the env vars to .env and restart -- the SDK detects them automatically.

Provider Key env vars
Anthropic (default) ANTHROPIC_API_KEY
OpenRouter ANTHROPIC_BASE_URL, OPENROUTER_API_KEY (see OpenRouter)
Vertex AI CLAUDE_CODE_USE_VERTEX=1, CLOUD_ML_REGION, ANTHROPIC_VERTEX_PROJECT_ID
Bedrock CLAUDE_CODE_USE_BEDROCK=1, AWS_REGION, AWS_ACCESS_KEY_ID, AWS_SECRET_ACCESS_KEY
Azure CLAUDE_CODE_USE_FOUNDRY=1, AZURE_FOUNDRY_RESOURCE, AZURE_API_KEY
Custom proxy ANTHROPIC_BASE_URL, ANTHROPIC_AUTH_TOKEN (optional)

API Reference

POST /query

Field Type Required Default Description
prompt string Yes -- The task for the agent (min 1 char)
anthropic_api_key string No $ANTHROPIC_API_KEY Anthropic key (falls back to env)
openrouter_api_key string No $OPENROUTER_API_KEY OpenRouter key (falls back to env)
e2b_api_key string No $E2B_API_KEY E2B key (falls back to env)
model string No from config Overrides sandstorm.json model
max_turns integer No from config Overrides sandstorm.json max_turns
timeout integer No 300 Sandbox lifetime in seconds
files object No null Files to upload ({path: content})

Response: text/event-stream

GET /health

Returns {"status": "ok"}

SSE Event Types

Type Description
system Session init -- tools, model, session ID
assistant Agent text + tool calls
user Tool execution results
result Final result with total_cost_usd, num_turns, and optional structured_output
error Error details (only on failure)

Client Examples

Python

import httpx
from httpx_sse import connect_sse

with httpx.Client() as client:
    with connect_sse(
        client, "POST",
        "https://your-sandstorm-host/query",
        json={
            "prompt": "Scrape the top 50 HN stories, cluster by topic, save to output/hn.csv"
        },
    ) as events:
        for sse in events.iter_sse():
            print(sse.data)

TypeScript

const res = await fetch("https://your-sandstorm-host/query", {
  method: "POST",
  headers: { "Content-Type": "application/json" },
  body: JSON.stringify({
    prompt: "Fetch recent arxiv papers on LLM agents, extract findings, write a lit review",
  }),
});

const reader = res.body!.getReader();
const decoder = new TextDecoder();
while (true) {
  const { done, value } = await reader.read();
  if (done) break;
  console.log(decoder.decode(value));
}

Deployment

Sandstorm is a stateless FastAPI app. Each request creates an independent E2B sandbox, runs the agent, and tears it down. No shared state, no sticky sessions, no coordination between requests. This means deploying for concurrent agent runs is trivial -- just add workers.

Production Server

For development or simple deployments, use the built-in server:

duvo-sandstorm serve --host 0.0.0.0 --port 8000

For production with multiple workers, use Gunicorn with uvicorn workers. Each worker handles multiple concurrent requests via async I/O:

pip install gunicorn
gunicorn sandstorm.main:app \
  --worker-class uvicorn.workers.UvicornWorker \
  --workers 4 \
  --bind 0.0.0.0:8000 \
  --timeout 600

Set --workers based on your machine (2× CPU cores is a reasonable starting point). Set --timeout higher than your longest expected agent run.

Running Many Agents Concurrently

Fire as many requests as you want. Each one gets its own sandbox:

import asyncio
import httpx
from httpx_sse import aconnect_sse


async def run_agent(client: httpx.AsyncClient, prompt: str):
    async with aconnect_sse(
        client, "POST", "http://localhost:8000/query",
        json={"prompt": prompt},
    ) as events:
        async for sse in events.aiter_sse():
            print(sse.data)


async def main():
    prompts = [
        "Scrape the top 50 YC companies and save as CSV",
        "Analyze Python dependency security for requests==2.31.0",
        "Fetch today's arxiv papers on LLM agents and write a summary",
        "Build a SQLite DB of US national parks from NPS.gov",
    ]
    async with httpx.AsyncClient(timeout=600) as client:
        await asyncio.gather(*[run_agent(client, p) for p in prompts])

asyncio.run(main())

All four agents run simultaneously in isolated sandboxes. They can't see each other. When one finishes, its VM is destroyed -- the others keep running.

Scaling

The Sandstorm server does almost no work itself -- it just proxies between your client and E2B. The real compute happens in E2B's cloud VMs. This means:

  • Horizontal scaling -- run multiple Sandstorm instances behind a load balancer. No shared state to worry about.
  • Bottleneck is E2B -- your concurrent sandbox limit depends on your E2B plan. The free tier allows a handful; paid plans scale higher.
  • CPU/memory on the server is minimal -- each request holds an open SSE connection and streams stdout. A single 2-core machine can comfortably handle dozens of concurrent agents.

Docker

FROM python:3.12-slim
WORKDIR /app
COPY . .
RUN pip install --no-cache-dir .  # or: pip install duvo-sandstorm
EXPOSE 8000
CMD ["duvo-sandstorm", "serve", "--host", "0.0.0.0", "--port", "8000"]
docker build -t sandstorm .
docker run -p 8000:8000 --env-file .env sandstorm

Deploy this container to any platform -- Railway, Fly.io, Cloud Run, ECS, Kubernetes. Since there's no state to persist, scaling up or down is just changing the replica count.

Vercel

One-click deploy:

Deploy with Vercel

The repo includes vercel.json and api/index.py pre-configured. Set ANTHROPIC_API_KEY and E2B_API_KEY as environment variables in your Vercel project settings.

Note: Vercel serverless functions have a maximum duration of 300s on Pro plans (10s on Hobby). For long-running agent tasks, use the Docker deployment or a dedicated server instead.

Security

  • Isolated execution -- every request gets a fresh VM sandbox, destroyed after
  • No server secrets -- keys via .env or per-request, never stored server-side
  • No shell injection -- prompts and config written as files, never interpolated into commands
  • Path traversal prevention -- file upload paths are normalized and validated
  • Structured errors -- failures stream as SSE error events, not silent drops
  • No persistence -- nothing survives between requests

Note: The Anthropic API key is passed into the sandbox as an environment variable (the SDK requires it). The agent runs with bypassPermissions mode, so it has full access to the sandbox environment. Use per-request keys with spending limits for untrusted callers.

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

duvo_sandstorm-0.3.0.tar.gz (21.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

duvo_sandstorm-0.3.0-py3-none-any.whl (19.8 kB view details)

Uploaded Python 3

File details

Details for the file duvo_sandstorm-0.3.0.tar.gz.

File metadata

  • Download URL: duvo_sandstorm-0.3.0.tar.gz
  • Upload date:
  • Size: 21.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.8.4

File hashes

Hashes for duvo_sandstorm-0.3.0.tar.gz
Algorithm Hash digest
SHA256 1ea721dd4c0f2ebc174c93f985591d6e32d618a2468482687ae60c22a10ea171
MD5 7c4b44ea171a5b96d54096487b137cd8
BLAKE2b-256 db31cedf9451c1a75eb862516d79046a099c6451dfb1dce332e289639938363f

See more details on using hashes here.

File details

Details for the file duvo_sandstorm-0.3.0-py3-none-any.whl.

File metadata

File hashes

Hashes for duvo_sandstorm-0.3.0-py3-none-any.whl
Algorithm Hash digest
SHA256 b1a78f1a82d8db202107a0d78f03a31ba90a9eec121b8750f7a61bbd7b1f3d5d
MD5 b8c6c32cabd8e6af2b97d45a44c8014d
BLAKE2b-256 6558ef6f496c8eb6526eb2f7b11c45c61ef9d2a033ff3b33df098a240f252412

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page