Universal MCP server for querying GraphQL, REST, and gRPC APIs using natural language
Project description
Ratatoskr
Turn any API into an MCP server. Query in English. Get results — even when the API can't.
Quick Start · Demos · How It Works · Providers · Reference · Development
Ratatoskr is a polyglot-LLM fork of agoda-com/api-agent — Agoda's universal API-to-MCP bridge. This fork adds first-class Anthropic and OpenAI-compatible (Ollama, LM Studio, vLLM) provider support alongside the original OpenAI backend. All credit for the core architecture goes to the Agoda engineering team.
Point at any GraphQL or REST API. Ask questions in natural language. The agent fetches data, stores it in DuckDB, and runs SQL post-processing. Rankings, filters, JOINs work even if the API doesn't support them.
What Makes It Different
Zero config. No custom MCP code per API. Point at a GraphQL endpoint or OpenAPI spec — schema introspected automatically.
SQL post-processing. API returns 10,000 unsorted rows? Agent ranks top 10. No GROUP BY? Agent aggregates. Need JOINs across endpoints? Agent combines.
Safe by default. Read-only. Mutations blocked unless explicitly allowed.
Recipe learning. Successful queries become cached pipelines. Reuse instantly without LLM reasoning.
Polyglot LLM. Run with OpenAI, Anthropic (Claude), or any OpenAI-compatible endpoint — same capabilities, your choice of model.
Quick Start
1. Run (choose one):
# OpenAI (default)
OPENAI_API_KEY=your_key uv run api-agent
# Anthropic (Claude)
uv run api-agent --provider anthropic --api-key your_key
# Local model (Ollama, LM Studio, vLLM)
uv run api-agent --provider openai-compat --base-url http://localhost:11434/v1 --model llama3
# Or Docker (OpenAI)
docker build -t ratatoskr .
docker run -p 3000:3000 -e OPENAI_API_KEY=your_key ratatoskr
2. Add to any MCP client:
{
"mcpServers": {
"rickandmorty": {
"url": "http://localhost:3000/mcp",
"headers": {
"X-Target-URL": "https://rickandmortyapi.com/graphql",
"X-API-Type": "graphql"
}
}
}
}
3. Ask questions:
- "Show characters from Earth, only alive ones, group by species"
- "Top 10 characters by episode count"
- "Compare alive vs dead by species, only species with 10+ characters"
That's it. Agent introspects schema, generates queries, runs SQL post-processing.
Try the Demos
Three public APIs included — just bring an LLM key:
# Set your LLM key (Anthropic by default, or override with API_AGENT_PROVIDER=openai)
export ANTHROPIC_API_KEY="your_key"
# Launch all three demo instances
./samples/run-demos.sh
Then connect MCP Inspector to any instance:
# Star Wars (GraphQL) — characters, films, planets, species, starships
npx @modelcontextprotocol/inspector --transport http --server-url http://localhost:3941/mcp
# Dad Jokes (GraphQL) — random jokes, search
npx @modelcontextprotocol/inspector --transport http --server-url http://localhost:3942/mcp
# NASA APOD (REST) — Astronomy Picture of the Day
npx @modelcontextprotocol/inspector --transport http --server-url http://localhost:3943/mcp
Sample questions to try:
| Demo | Try asking... |
|---|---|
| Star Wars | "List all films with their directors, sorted by release date" |
| Star Wars | "Which planet has the most characters? Show top 5" |
| Dad Jokes | "Find me jokes about cats" |
| NASA APOD | "What was the astronomy picture on 2024-01-01?" |
More Examples
REST API (Petstore):
{
"mcpServers": {
"petstore": {
"url": "http://localhost:3000/mcp",
"headers": {
"X-Target-URL": "https://petstore3.swagger.io/api/v3/openapi.json",
"X-API-Type": "rest"
}
}
}
}
Your own API with auth:
{
"mcpServers": {
"myapi": {
"url": "http://localhost:3000/mcp",
"headers": {
"X-Target-URL": "https://api.example.com/graphql",
"X-API-Type": "graphql",
"X-Target-Headers": "{\"Authorization\": \"Bearer YOUR_TOKEN\"}"
}
}
}
}
How It Works
sequenceDiagram
participant U as User
participant M as MCP Server
participant A as Agent
participant G as Target API
U->>M: Question + Headers
M->>G: Schema introspection
G-->>M: Schema
M->>A: Schema + question
A->>G: API call
G-->>A: Data stored in DuckDB
A->>A: SQL post-processing
A-->>M: Summary
M-->>U: {ok, data, queries[]}
Architecture
flowchart TB
subgraph Client["MCP Client"]
H["Headers: X-Target-URL, X-API-Type"]
end
subgraph MCP["MCP Server (FastMCP)"]
Q["{prefix}_query"]
E["{prefix}_execute"]
R["r_{recipe} (dynamic)"]
end
subgraph Agent["Agents (Polyglot LLM)"]
GA["GraphQL Agent"]
RA["REST Agent"]
end
subgraph Exec["Executors"]
HTTP["HTTP Client"]
Duck["DuckDB"]
end
Client -->|NL + headers| MCP
Q -->|graphql| GA
Q -->|rest| RA
E --> HTTP
R -->|"no LLM"| HTTP
R --> Duck
GA --> HTTP
RA --> HTTP
GA --> Duck
RA --> Duck
HTTP --> API[Target API]
Stack: FastMCP · OpenAI / Anthropic / OpenAI-compatible · DuckDB
Token-Optimized Output (TOON)
API responses are automatically compressed using TOON format before being sent to the LLM. TOON strips JSON's structural punctuation (braces, quotes, colons) that creates noise tokens diluting LLM attention, concentrating the model's focus on actual field names and values. Typical reduction: 30-60% fewer tokens with improved reasoning quality.
- Default-on for tool results and SQL query output
- Graceful JSON fallback if TOON produces larger output
- Disable globally:
API_AGENT_TOON_TOOL_RESULTS_ENABLED=false
Recipe Learning
Agent learns reusable patterns from successful queries:
- Executes — API calls + SQL via LLM reasoning
- Extracts — LLM converts trace into parameterized template
- Caches — Stores recipe keyed by (API, schema hash)
- Exposes — Recipe becomes MCP tool (
r_{name}) callable without LLM
flowchart LR
subgraph First["First Query via {prefix}_query"]
Q1["'Top 5 users by age'"]
A1["Agent reasons"]
E1["API + SQL"]
R1["Recipe extracted"]
end
subgraph Tools["MCP Tools"]
T["r_get_top_users<br/>params: {limit}"]
end
subgraph Reuse["Direct Call"]
Q2["r_get_top_users({limit: 10})"]
X["Execute directly"]
end
Q1 --> A1 --> E1 --> R1 --> T
Q2 --> T --> X
Recipes auto-expire on schema changes. Disable with API_AGENT_ENABLE_RECIPES=false.
Providers
Ratatoskr supports multiple LLM providers through a thin abstraction layer.
OpenAI (default)
OPENAI_API_KEY=sk-... uv run api-agent
Anthropic (Claude)
# Via CLI
uv run api-agent --provider anthropic --api-key sk-ant-...
# Via env vars
API_AGENT_PROVIDER=anthropic ANTHROPIC_API_KEY=sk-ant-... uv run api-agent
# Custom model
uv run api-agent --provider anthropic --model claude-opus-4-20250514
Local Models (Ollama, LM Studio, vLLM)
# Ollama
uv run api-agent --provider openai-compat \
--base-url http://localhost:11434/v1 \
--model llama3
# LM Studio
uv run api-agent --provider openai-compat \
--base-url http://localhost:1234/v1 \
--model local-model
# vLLM
uv run api-agent --provider openai-compat \
--base-url http://gpu-server:8000/v1 \
--model mistral-7b
Note: Local models must support tool/function calling for full functionality. If an endpoint doesn't support tools, the agent will retry without them (graceful degradation).
Reference
Headers
| Header | Required | Description |
|---|---|---|
X-Target-URL |
Yes | GraphQL endpoint OR OpenAPI spec URL |
X-API-Type |
Yes | graphql or rest |
X-Target-Headers |
No | JSON auth headers, e.g. {"Authorization": "Bearer xxx"} |
X-API-Name |
No | Override tool name prefix (default: auto-generated) |
X-Base-URL |
No | Override base URL for REST API calls |
X-Allow-Unsafe-Paths |
No | Header string containing JSON array of fnmatch globs (*, ?) for POST/PUT/DELETE/PATCH |
X-Poll-Paths |
No | Header string containing JSON array of polling path patterns (enables poll tool) |
X-Include-Result |
No | Include full uncapped result field in output |
X-Allow-Endpoints |
No | JSON array of glob patterns to restrict exposed endpoints |
Header value examples
X-Allow-Unsafe-Paths and X-Poll-Paths use the same escaping format: JSON array encoded as a header string.
MCP config (JSON):
{
"headers": {
"X-Allow-Unsafe-Paths": "[\"/search\", \"/api/*/query\", \"/jobs/*/cancel\"]",
"X-Poll-Paths": "[\"/search\", \"/trips/*/status\"]"
}
}
X-Allow-Unsafe-Paths pattern examples:
"/search"exact path"/api/*/query"one wildcard segment"/jobs/*"any suffix under/jobs/
X-Poll-Paths pattern examples:
"/search"exact polling path"/trips/*/status"wildcard polling path
X-Poll-Paths enables polling guidance/tooling; X-Allow-Unsafe-Paths controls unsafe method allowlist.
Escaping quick check (same for both headers):
- wrong:
"X-Allow-Unsafe-Paths": "["/search"]" - right:
"X-Allow-Unsafe-Paths": "[\"/search\"]"
MCP Tools
Core tools (2 per API):
| Tool | Input | Output |
|---|---|---|
{prefix}_query |
Natural language question | {ok, data, queries/api_calls} |
{prefix}_execute |
GraphQL: query, variables / REST: method, path, params |
{ok, data} |
Tool names auto-generated from URL (e.g., example_query). Override with X-API-Name.
Recipe tools (dynamic, added as recipes are learned):
| Tool | Input | Output |
|---|---|---|
r_{recipe_slug} |
flat recipe-specific params, return_directly (bool) |
CSV or {ok, data, executed_queries/calls} |
Cached pipelines, no LLM reasoning. Appear after successful queries. Clients notified via tools/list_changed.
CLI Arguments
| Argument | Description |
|---|---|
--provider |
LLM provider: openai, anthropic, or openai-compat |
--model |
Model name (default: provider-specific) |
--api-key |
API key (overrides env vars) |
--base-url |
Custom LLM endpoint (required for openai-compat) |
--port |
Server port (default: 3000) |
--host |
Server host (default: 0.0.0.0) |
--transport |
MCP transport: http, streamable-http, sse |
--profile |
Config profile: local (relaxes security for local dev) |
--debug |
Enable debug logging |
CLI arguments override environment variables.
Configuration (env vars)
| Variable | Required | Default | Description |
|---|---|---|---|
API_AGENT_PROVIDER |
No | openai |
LLM provider (openai, anthropic, openai-compat) |
API_AGENT_API_KEY |
Yes | - | API key (also accepts OPENAI_API_KEY, ANTHROPIC_API_KEY) |
API_AGENT_BASE_URL |
No* | - | Custom LLM endpoint (*required for openai-compat) |
API_AGENT_MODEL_NAME |
No | (provider default) | Model name |
API_AGENT_PORT |
No | 3000 | Server port |
API_AGENT_ENABLE_RECIPES |
No | true | Enable recipe learning & caching |
API_AGENT_RECIPE_CACHE_SIZE |
No | 64 | Max cached recipes (LRU eviction) |
API_AGENT_ALLOW_ENDPOINTS_REST |
No | - | CSV glob patterns for REST endpoint allowlist |
API_AGENT_ALLOW_ENDPOINTS_GRAPHQL |
No | - | CSV glob patterns for GraphQL endpoint allowlist |
API_AGENT_ALLOW_ENDPOINTS_GRPC |
No | - | CSV glob patterns for gRPC endpoint allowlist |
API_AGENT_PROFILE |
No | - | Config profile (local = relax for local dev) |
API_AGENT_SCHEMA_REDUCTION_PROVIDER |
No | (inherits PROVIDER) |
LLM provider for schema reduction |
API_AGENT_SCHEMA_REDUCTION_MODEL |
No | (provider default) | Model for schema reduction |
API_AGENT_SCHEMA_REDUCTION_API_KEY |
No | (inherits API_KEY) |
API key for schema reduction LLM |
API_AGENT_SCHEMA_REDUCTION_BASE_URL |
No | (inherits BASE_URL) |
Endpoint for schema reduction LLM |
OTEL_EXPORTER_OTLP_ENDPOINT |
No | - | OpenTelemetry tracing endpoint |
Provider defaults:
| Provider | Default model | API key env var |
|---|---|---|
openai |
gpt-4o |
OPENAI_API_KEY |
anthropic |
claude-sonnet-4-20250514 |
ANTHROPIC_API_KEY |
openai-compat |
gpt-4o |
(optional) |
Local development
Use PROFILE=local (or --profile local) to set sensible defaults for local dev:
# All three of these are set automatically:
# BLOCK_PRIVATE_IPS=false (allow localhost targets)
# LOG_FORMAT=console (human-readable logs)
# SCHEMA_REDUCTION_ENABLED=false (no cloud key needed)
uv run api-agent --profile local \
--provider openai-compat \
--base-url http://localhost:11434/v1 \
--model llama3
Explicit env vars always override profile defaults (e.g., BLOCK_PRIVATE_IPS=true wins even with PROFILE=local).
Endpoint Allowlisting
Large APIs (500+ endpoints) can overwhelm LLM context. Endpoint allowlisting filters schemas before the LLM sees them, so agents only operate on permitted endpoints.
Config (ops ceiling)
Set per-protocol env vars with comma-separated fnmatch glob patterns:
API_AGENT_ALLOW_ENDPOINTS_REST="GET /users/*,GET /accounts/*"
API_AGENT_ALLOW_ENDPOINTS_GRAPHQL="Query.users*,Query.accounts*"
API_AGENT_ALLOW_ENDPOINTS_GRPC="myapp.UserService/*,myapp.AccountService/*"
Per-session header (narrows config)
Clients send X-Allow-Endpoints as a JSON array of glob patterns:
{ "X-Allow-Endpoints": "[\"GET /users/*\"]" }
Intersection semantics: When both config and header are set, an endpoint must match a pattern from each. The header can only narrow the config ceiling, never widen it.
Match target format
| Protocol | Format | Examples |
|---|---|---|
| REST | METHOD /path |
GET /users/*, POST /search, * /api/v2/* |
| GraphQL | Query.fieldName |
Query.users, Query.account* |
| gRPC | package.Service/Method |
myapp.UserService/*, myapp.*/Get* |
Behavior
- No config + no header = all endpoints exposed (default)
- Allowlist active, some match = agent sees only matching endpoints
- Allowlist active, none match = clear error returned (agent does not run)
search_schema()operates on the filtered schema — blocked endpoints are invisible
Roadmap
Planned improvements (contributions welcome):
- Streaming responses — Stream agent reasoning and partial results to MCP clients
- Mutation support — Controlled write operations with confirmation flows
- Schema caching — Cache introspected schemas to reduce startup latency
- Multi-API joins — Query across multiple APIs in a single request
- Recipe sharing — Export/import learned recipes between instances
- WebSocket subscriptions — Support GraphQL subscriptions for real-time data
- Plugin system — Custom pre/post-processing hooks for API responses
Development
git clone https://github.com/innago-property-management/ratatoskr.git
cd ratatoskr
uv sync --group dev
uv run pytest tests/ -v # Tests (1236 passing)
uv run ruff check api_agent/ # Lint
uv run ty check # Type check
Kubernetes
Kustomize manifests are in deploy/:
deploy/
base/ # Deployment + Service (1 replica)
overlays/
production/ # 2 replicas, PDB, topology spread
Requirements: Kubernetes >= 1.21 (production overlay uses policy/v1 PodDisruptionBudget).
Probes:
/health— liveness (process alive)/ready— readiness (config valid, API key present for cloud providers;openai-compatproviders like Ollama/vLLM don't require a key)
Observability
Set OTEL_EXPORTER_OTLP_ENDPOINT to enable OpenTelemetry tracing. Works with Jaeger, Zipkin, Grafana Tempo, Arize Phoenix.
Origin & Attribution
Ratatoskr is a fork of api-agent by Agoda, licensed under the MIT License.
The core architecture — FastMCP server, dynamic tool naming, agent orchestration, DuckDB post-processing, and recipe learning — is entirely Agoda's work. Ratatoskr extends it with:
- Polyglot LLM support — Anthropic, OpenAI, and OpenAI-compatible providers via a pluggable
LLMProviderabstraction - Token-Optimized Output (TOON) — Strips JSON punctuation noise to improve LLM attention quality with 30-60% fewer tokens
- Schema reduction — 3-layer pipeline (keyword ranking, TOON, AI) using any configured LLM provider, not just Anthropic
- Local dev profile —
PROFILE=localcollapses three manual overrides into one env var - Expanded test coverage — 1412 tests covering orchestration, safety boundaries, configuration contracts, and provider SDK surfaces
- GraphQL partial success fix — Returns both
dataanderrorswhen both present, per the GraphQL specification
The name Ratatoskr comes from the Norse squirrel who runs up and down Yggdrasil carrying messages between realms — a fitting metaphor for a universal API-to-LLM bridge.
Upstream: agoda-com/api-agent · Blog post
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file api_agent_ratatoskr-2.0.0.tar.gz.
File metadata
- Download URL: api_agent_ratatoskr-2.0.0.tar.gz
- Upload date:
- Size: 2.1 MB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
ae87aa6d0504c7cdcb7f1c1bf32087ca97f664a2e0bf0162568efaefdd11b41c
|
|
| MD5 |
27e1bdbd4d7c56f3bfdb4e540d9a8e1c
|
|
| BLAKE2b-256 |
9320bd6360652660a54bca95d4b0401b0679c2adc983c3320b8a34aeb0bf1f7e
|
Provenance
The following attestation bundles were made for api_agent_ratatoskr-2.0.0.tar.gz:
Publisher:
release.yml on innago-property-management/ratatoskr
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
api_agent_ratatoskr-2.0.0.tar.gz -
Subject digest:
ae87aa6d0504c7cdcb7f1c1bf32087ca97f664a2e0bf0162568efaefdd11b41c - Sigstore transparency entry: 1200728599
- Sigstore integration time:
-
Permalink:
innago-property-management/ratatoskr@49add90efa02e153d8aa11efb323d052b8a0db9e -
Branch / Tag:
refs/tags/2.0.0 - Owner: https://github.com/innago-property-management
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
release.yml@49add90efa02e153d8aa11efb323d052b8a0db9e -
Trigger Event:
release
-
Statement type:
File details
Details for the file api_agent_ratatoskr-2.0.0-py3-none-any.whl.
File metadata
- Download URL: api_agent_ratatoskr-2.0.0-py3-none-any.whl
- Upload date:
- Size: 152.0 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
fc8b81496b9872a7eb1d491a413636fd0020cdb4ce4fd0bd8dd0ef352fd34a32
|
|
| MD5 |
c1a962a84bb978998cc330269a8fd8ac
|
|
| BLAKE2b-256 |
c8b08afc123d449c15148b9872cea8dccb703f53a10fc5493b92f57eb54a22eb
|
Provenance
The following attestation bundles were made for api_agent_ratatoskr-2.0.0-py3-none-any.whl:
Publisher:
release.yml on innago-property-management/ratatoskr
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
api_agent_ratatoskr-2.0.0-py3-none-any.whl -
Subject digest:
fc8b81496b9872a7eb1d491a413636fd0020cdb4ce4fd0bd8dd0ef352fd34a32 - Sigstore transparency entry: 1200728607
- Sigstore integration time:
-
Permalink:
innago-property-management/ratatoskr@49add90efa02e153d8aa11efb323d052b8a0db9e -
Branch / Tag:
refs/tags/2.0.0 - Owner: https://github.com/innago-property-management
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
release.yml@49add90efa02e153d8aa11efb323d052b8a0db9e -
Trigger Event:
release
-
Statement type: