Skip to main content

Universal MCP server for querying GraphQL, REST, and gRPC APIs using natural language

Project description

Ratatoskr

Ratatoskr

Turn any API into an MCP server. Query in English. Get results — even when the API can't.

Quick Start · Demos · How It Works · Providers · Reference · Development


Ratatoskr is a polyglot-LLM fork of agoda-com/api-agent — Agoda's universal API-to-MCP bridge. This fork adds first-class Anthropic and OpenAI-compatible (Ollama, LM Studio, vLLM) provider support alongside the original OpenAI backend. All credit for the core architecture goes to the Agoda engineering team.


Point at any GraphQL or REST API. Ask questions in natural language. The agent fetches data, stores it in DuckDB, and runs SQL post-processing. Rankings, filters, JOINs work even if the API doesn't support them.

What Makes It Different

Zero config. No custom MCP code per API. Point at a GraphQL endpoint or OpenAPI spec — schema introspected automatically.

SQL post-processing. API returns 10,000 unsorted rows? Agent ranks top 10. No GROUP BY? Agent aggregates. Need JOINs across endpoints? Agent combines.

Safe by default. Read-only. Mutations blocked unless explicitly allowed.

Recipe learning. Successful queries become cached pipelines. Reuse instantly without LLM reasoning.

Polyglot LLM. Run with OpenAI, Anthropic (Claude), or any OpenAI-compatible endpoint — same capabilities, your choice of model.

Quick Start

1. Run (choose one):

# OpenAI (default)
OPENAI_API_KEY=your_key uv run api-agent

# Anthropic (Claude)
uv run api-agent --provider anthropic --api-key your_key

# Local model (Ollama, LM Studio, vLLM)
uv run api-agent --provider openai-compat --base-url http://localhost:11434/v1 --model llama3

# Or Docker (OpenAI)
docker build -t ratatoskr .
docker run -p 3000:3000 -e OPENAI_API_KEY=your_key ratatoskr

2. Add to any MCP client:

{
  "mcpServers": {
    "rickandmorty": {
      "url": "http://localhost:3000/mcp",
      "headers": {
        "X-Target-URL": "https://rickandmortyapi.com/graphql",
        "X-API-Type": "graphql"
      }
    }
  }
}

3. Ask questions:

  • "Show characters from Earth, only alive ones, group by species"
  • "Top 10 characters by episode count"
  • "Compare alive vs dead by species, only species with 10+ characters"

That's it. Agent introspects schema, generates queries, runs SQL post-processing.

Try the Demos

Three public APIs included — just bring an LLM key:

# Set your LLM key (Anthropic by default, or override with API_AGENT_PROVIDER=openai)
export ANTHROPIC_API_KEY="your_key"

# Launch all three demo instances
./samples/run-demos.sh

Then connect MCP Inspector to any instance:

# Star Wars (GraphQL) — characters, films, planets, species, starships
npx @modelcontextprotocol/inspector --transport http --server-url http://localhost:3941/mcp

# Dad Jokes (GraphQL) — random jokes, search
npx @modelcontextprotocol/inspector --transport http --server-url http://localhost:3942/mcp

# NASA APOD (REST) — Astronomy Picture of the Day
npx @modelcontextprotocol/inspector --transport http --server-url http://localhost:3943/mcp

Sample questions to try:

Demo Try asking...
Star Wars "List all films with their directors, sorted by release date"
Star Wars "Which planet has the most characters? Show top 5"
Dad Jokes "Find me jokes about cats"
NASA APOD "What was the astronomy picture on 2024-01-01?"

More Examples

REST API (Petstore):

{
  "mcpServers": {
    "petstore": {
      "url": "http://localhost:3000/mcp",
      "headers": {
        "X-Target-URL": "https://petstore3.swagger.io/api/v3/openapi.json",
        "X-API-Type": "rest"
      }
    }
  }
}

Your own API with auth:

{
  "mcpServers": {
    "myapi": {
      "url": "http://localhost:3000/mcp",
      "headers": {
        "X-Target-URL": "https://api.example.com/graphql",
        "X-API-Type": "graphql",
        "X-Target-Headers": "{\"Authorization\": \"Bearer YOUR_TOKEN\"}"
      }
    }
  }
}

How It Works

sequenceDiagram
    participant U as User
    participant M as MCP Server
    participant A as Agent
    participant G as Target API

    U->>M: Question + Headers
    M->>G: Schema introspection
    G-->>M: Schema
    M->>A: Schema + question
    A->>G: API call
    G-->>A: Data stored in DuckDB
    A->>A: SQL post-processing
    A-->>M: Summary
    M-->>U: {ok, data, queries[]}

Architecture

flowchart TB
    subgraph Client["MCP Client"]
        H["Headers: X-Target-URL, X-API-Type"]
    end

    subgraph MCP["MCP Server (FastMCP)"]
        Q["{prefix}_query"]
        E["{prefix}_execute"]
        R["r_{recipe} (dynamic)"]
    end

    subgraph Agent["Agents (Polyglot LLM)"]
        GA["GraphQL Agent"]
        RA["REST Agent"]
    end

    subgraph Exec["Executors"]
        HTTP["HTTP Client"]
        Duck["DuckDB"]
    end

    Client -->|NL + headers| MCP
    Q -->|graphql| GA
    Q -->|rest| RA
    E --> HTTP
    R -->|"no LLM"| HTTP
    R --> Duck
    GA --> HTTP
    RA --> HTTP
    GA --> Duck
    RA --> Duck
    HTTP --> API[Target API]

Stack: FastMCP · OpenAI / Anthropic / OpenAI-compatible · DuckDB


Token-Optimized Output (TOON)

API responses are automatically compressed using TOON format before being sent to the LLM. TOON strips JSON's structural punctuation (braces, quotes, colons) that creates noise tokens diluting LLM attention, concentrating the model's focus on actual field names and values. Typical reduction: 30-60% fewer tokens with improved reasoning quality.

  • Default-on for tool results and SQL query output
  • Graceful JSON fallback if TOON produces larger output
  • Disable globally: API_AGENT_TOON_TOOL_RESULTS_ENABLED=false

Recipe Learning

Agent learns reusable patterns from successful queries:

  1. Executes — API calls + SQL via LLM reasoning
  2. Extracts — LLM converts trace into parameterized template
  3. Caches — Stores recipe keyed by (API, schema hash)
  4. Exposes — Recipe becomes MCP tool (r_{name}) callable without LLM
flowchart LR
    subgraph First["First Query via {prefix}_query"]
        Q1["'Top 5 users by age'"]
        A1["Agent reasons"]
        E1["API + SQL"]
        R1["Recipe extracted"]
    end

    subgraph Tools["MCP Tools"]
        T["r_get_top_users<br/>params: {limit}"]
    end

    subgraph Reuse["Direct Call"]
        Q2["r_get_top_users({limit: 10})"]
        X["Execute directly"]
    end

    Q1 --> A1 --> E1 --> R1 --> T
    Q2 --> T --> X

Recipes auto-expire on schema changes. Disable with API_AGENT_ENABLE_RECIPES=false.


Providers

Ratatoskr supports multiple LLM providers through a thin abstraction layer.

OpenAI (default)

OPENAI_API_KEY=sk-... uv run api-agent

Anthropic (Claude)

# Via CLI
uv run api-agent --provider anthropic --api-key sk-ant-...

# Via env vars
API_AGENT_PROVIDER=anthropic ANTHROPIC_API_KEY=sk-ant-... uv run api-agent

# Custom model
uv run api-agent --provider anthropic --model claude-opus-4-20250514

Local Models (Ollama, LM Studio, vLLM)

# Ollama
uv run api-agent --provider openai-compat \
  --base-url http://localhost:11434/v1 \
  --model llama3

# LM Studio
uv run api-agent --provider openai-compat \
  --base-url http://localhost:1234/v1 \
  --model local-model

# vLLM
uv run api-agent --provider openai-compat \
  --base-url http://gpu-server:8000/v1 \
  --model mistral-7b

Note: Local models must support tool/function calling for full functionality. If an endpoint doesn't support tools, the agent will retry without them (graceful degradation).


Reference

Headers

Header Required Description
X-Target-URL Yes GraphQL endpoint OR OpenAPI spec URL
X-API-Type Yes graphql or rest
X-Target-Headers No JSON auth headers, e.g. {"Authorization": "Bearer xxx"}
X-API-Name No Override tool name prefix (default: auto-generated)
X-Base-URL No Override base URL for REST API calls
X-Allow-Unsafe-Paths No Header string containing JSON array of fnmatch globs (*, ?) for POST/PUT/DELETE/PATCH
X-Poll-Paths No Header string containing JSON array of polling path patterns (enables poll tool)
X-Include-Result No Include full uncapped result field in output
X-Allow-Endpoints No JSON array of glob patterns to restrict exposed endpoints

Header value examples

X-Allow-Unsafe-Paths and X-Poll-Paths use the same escaping format: JSON array encoded as a header string.

MCP config (JSON):

{
  "headers": {
    "X-Allow-Unsafe-Paths": "[\"/search\", \"/api/*/query\", \"/jobs/*/cancel\"]",
    "X-Poll-Paths": "[\"/search\", \"/trips/*/status\"]"
  }
}

X-Allow-Unsafe-Paths pattern examples:

  • "/search" exact path
  • "/api/*/query" one wildcard segment
  • "/jobs/*" any suffix under /jobs/

X-Poll-Paths pattern examples:

  • "/search" exact polling path
  • "/trips/*/status" wildcard polling path

X-Poll-Paths enables polling guidance/tooling; X-Allow-Unsafe-Paths controls unsafe method allowlist.

Escaping quick check (same for both headers):

  • wrong: "X-Allow-Unsafe-Paths": "["/search"]"
  • right: "X-Allow-Unsafe-Paths": "[\"/search\"]"

MCP Tools

Core tools (2 per API):

Tool Input Output
{prefix}_query Natural language question {ok, data, queries/api_calls}
{prefix}_execute GraphQL: query, variables / REST: method, path, params {ok, data}

Tool names auto-generated from URL (e.g., example_query). Override with X-API-Name.

Recipe tools (dynamic, added as recipes are learned):

Tool Input Output
r_{recipe_slug} flat recipe-specific params, return_directly (bool) CSV or {ok, data, executed_queries/calls}

Cached pipelines, no LLM reasoning. Appear after successful queries. Clients notified via tools/list_changed.

CLI Arguments

Argument Description
--provider LLM provider: openai, anthropic, or openai-compat
--model Model name (default: provider-specific)
--api-key API key (overrides env vars)
--base-url Custom LLM endpoint (required for openai-compat)
--port Server port (default: 3000)
--host Server host (default: 0.0.0.0)
--transport MCP transport: http, streamable-http, sse
--profile Config profile: local (relaxes security for local dev)
--debug Enable debug logging

CLI arguments override environment variables.

Configuration (env vars)

Variable Required Default Description
API_AGENT_PROVIDER No openai LLM provider (openai, anthropic, openai-compat)
API_AGENT_API_KEY Yes - API key (also accepts OPENAI_API_KEY, ANTHROPIC_API_KEY)
API_AGENT_BASE_URL No* - Custom LLM endpoint (*required for openai-compat)
API_AGENT_MODEL_NAME No (provider default) Model name
API_AGENT_PORT No 3000 Server port
API_AGENT_ENABLE_RECIPES No true Enable recipe learning & caching
API_AGENT_RECIPE_CACHE_SIZE No 64 Max cached recipes (LRU eviction)
API_AGENT_ALLOW_ENDPOINTS_REST No - CSV glob patterns for REST endpoint allowlist
API_AGENT_ALLOW_ENDPOINTS_GRAPHQL No - CSV glob patterns for GraphQL endpoint allowlist
API_AGENT_ALLOW_ENDPOINTS_GRPC No - CSV glob patterns for gRPC endpoint allowlist
API_AGENT_PROFILE No - Config profile (local = relax for local dev)
API_AGENT_SCHEMA_REDUCTION_PROVIDER No (inherits PROVIDER) LLM provider for schema reduction
API_AGENT_SCHEMA_REDUCTION_MODEL No (provider default) Model for schema reduction
API_AGENT_SCHEMA_REDUCTION_API_KEY No (inherits API_KEY) API key for schema reduction LLM
API_AGENT_SCHEMA_REDUCTION_BASE_URL No (inherits BASE_URL) Endpoint for schema reduction LLM
OTEL_EXPORTER_OTLP_ENDPOINT No - OpenTelemetry tracing endpoint

Provider defaults:

Provider Default model API key env var
openai gpt-4o OPENAI_API_KEY
anthropic claude-sonnet-4-20250514 ANTHROPIC_API_KEY
openai-compat gpt-4o (optional)

Local development

Use PROFILE=local (or --profile local) to set sensible defaults for local dev:

# All three of these are set automatically:
#   BLOCK_PRIVATE_IPS=false    (allow localhost targets)
#   LOG_FORMAT=console         (human-readable logs)
#   SCHEMA_REDUCTION_ENABLED=false  (no cloud key needed)

uv run api-agent --profile local \
  --provider openai-compat \
  --base-url http://localhost:11434/v1 \
  --model llama3

Explicit env vars always override profile defaults (e.g., BLOCK_PRIVATE_IPS=true wins even with PROFILE=local).


Endpoint Allowlisting

Large APIs (500+ endpoints) can overwhelm LLM context. Endpoint allowlisting filters schemas before the LLM sees them, so agents only operate on permitted endpoints.

Config (ops ceiling)

Set per-protocol env vars with comma-separated fnmatch glob patterns:

API_AGENT_ALLOW_ENDPOINTS_REST="GET /users/*,GET /accounts/*"
API_AGENT_ALLOW_ENDPOINTS_GRAPHQL="Query.users*,Query.accounts*"
API_AGENT_ALLOW_ENDPOINTS_GRPC="myapp.UserService/*,myapp.AccountService/*"

Per-session header (narrows config)

Clients send X-Allow-Endpoints as a JSON array of glob patterns:

{ "X-Allow-Endpoints": "[\"GET /users/*\"]" }

Intersection semantics: When both config and header are set, an endpoint must match a pattern from each. The header can only narrow the config ceiling, never widen it.

Match target format

Protocol Format Examples
REST METHOD /path GET /users/*, POST /search, * /api/v2/*
GraphQL Query.fieldName Query.users, Query.account*
gRPC package.Service/Method myapp.UserService/*, myapp.*/Get*

Behavior

  • No config + no header = all endpoints exposed (default)
  • Allowlist active, some match = agent sees only matching endpoints
  • Allowlist active, none match = clear error returned (agent does not run)
  • search_schema() operates on the filtered schema — blocked endpoints are invisible

Roadmap

Planned improvements (contributions welcome):

  • Streaming responses — Stream agent reasoning and partial results to MCP clients
  • Mutation support — Controlled write operations with confirmation flows
  • Schema caching — Cache introspected schemas to reduce startup latency
  • Multi-API joins — Query across multiple APIs in a single request
  • Recipe sharing — Export/import learned recipes between instances
  • WebSocket subscriptions — Support GraphQL subscriptions for real-time data
  • Plugin system — Custom pre/post-processing hooks for API responses

Development

git clone https://github.com/innago-property-management/ratatoskr.git
cd ratatoskr
uv sync --group dev
uv run pytest tests/ -v      # Tests (1236 passing)
uv run ruff check api_agent/  # Lint
uv run ty check               # Type check

Kubernetes

Kustomize manifests are in deploy/:

deploy/
  base/          # Deployment + Service (1 replica)
  overlays/
    production/  # 2 replicas, PDB, topology spread

Requirements: Kubernetes >= 1.21 (production overlay uses policy/v1 PodDisruptionBudget).

Probes:

  • /health — liveness (process alive)
  • /ready — readiness (config valid, API key present for cloud providers; openai-compat providers like Ollama/vLLM don't require a key)

Observability

Set OTEL_EXPORTER_OTLP_ENDPOINT to enable OpenTelemetry tracing. Works with Jaeger, Zipkin, Grafana Tempo, Arize Phoenix.


Origin & Attribution

Ratatoskr is a fork of api-agent by Agoda, licensed under the MIT License.

The core architecture — FastMCP server, dynamic tool naming, agent orchestration, DuckDB post-processing, and recipe learning — is entirely Agoda's work. Ratatoskr extends it with:

  • Polyglot LLM support — Anthropic, OpenAI, and OpenAI-compatible providers via a pluggable LLMProvider abstraction
  • Token-Optimized Output (TOON) — Strips JSON punctuation noise to improve LLM attention quality with 30-60% fewer tokens
  • Schema reduction — 3-layer pipeline (keyword ranking, TOON, AI) using any configured LLM provider, not just Anthropic
  • Local dev profilePROFILE=local collapses three manual overrides into one env var
  • Expanded test coverage — 1412 tests covering orchestration, safety boundaries, configuration contracts, and provider SDK surfaces
  • GraphQL partial success fix — Returns both data and errors when both present, per the GraphQL specification

The name Ratatoskr comes from the Norse squirrel who runs up and down Yggdrasil carrying messages between realms — a fitting metaphor for a universal API-to-LLM bridge.

Upstream: agoda-com/api-agent · Blog post

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

api_agent_ratatoskr-2.0.0.tar.gz (2.1 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

api_agent_ratatoskr-2.0.0-py3-none-any.whl (152.0 kB view details)

Uploaded Python 3

File details

Details for the file api_agent_ratatoskr-2.0.0.tar.gz.

File metadata

  • Download URL: api_agent_ratatoskr-2.0.0.tar.gz
  • Upload date:
  • Size: 2.1 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for api_agent_ratatoskr-2.0.0.tar.gz
Algorithm Hash digest
SHA256 ae87aa6d0504c7cdcb7f1c1bf32087ca97f664a2e0bf0162568efaefdd11b41c
MD5 27e1bdbd4d7c56f3bfdb4e540d9a8e1c
BLAKE2b-256 9320bd6360652660a54bca95d4b0401b0679c2adc983c3320b8a34aeb0bf1f7e

See more details on using hashes here.

Provenance

The following attestation bundles were made for api_agent_ratatoskr-2.0.0.tar.gz:

Publisher: release.yml on innago-property-management/ratatoskr

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file api_agent_ratatoskr-2.0.0-py3-none-any.whl.

File metadata

File hashes

Hashes for api_agent_ratatoskr-2.0.0-py3-none-any.whl
Algorithm Hash digest
SHA256 fc8b81496b9872a7eb1d491a413636fd0020cdb4ce4fd0bd8dd0ef352fd34a32
MD5 c1a962a84bb978998cc330269a8fd8ac
BLAKE2b-256 c8b08afc123d449c15148b9872cea8dccb703f53a10fc5493b92f57eb54a22eb

See more details on using hashes here.

Provenance

The following attestation bundles were made for api_agent_ratatoskr-2.0.0-py3-none-any.whl:

Publisher: release.yml on innago-property-management/ratatoskr

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page