Skip to main content

Free-AI gateway: OpenAI-compatible local proxy that orchestrates free-tier inference across multiple providers

Project description

FreeRide

One free AI endpoint. Five providers behind it. Your agents don't need to know.

$ curl -sSL https://free-ride.xyz/install.sh | sh
$ export OPENROUTER_API_KEY=sk-or-v1-...
$ freeride serve

freeride gateway listening on http://127.0.0.1:11343
  providers: openrouter, groq, huggingface
  point any OpenAI-compatible agent at:
    OPENAI_API_BASE=http://127.0.0.1:11343/v1
    OPENAI_API_KEY=any

That's it. Aider, Continue, OpenClaw, Hermes, the OpenAI Python SDK — anything that speaks OpenAI now speaks every free tier you have a key for.

Demo

┌─ your agent ─────────┐         ┌─ freeride (localhost) ─┐         ┌─ providers ─┐
│                      │  POST   │                        │         │             │
│  chat.completions    │────────▶│  pick provider         │────────▶│  OpenRouter │ 429
│   .create(...)       │         │  pick key (not cooling)│  retry  │     ↓       │
│                      │         │  forward request       │────────▶│  Groq       │ ✓
│  ◀───────────────────│   200   │  ◀─────────────────────│         │             │
│                      │         │                        │         │  NIM, CF,   │
│                      │         │  X-FreeRide-Provider:  │         │  HF — only  │
│                      │         │   groq                 │         │  if needed  │
└──────────────────────┘         └────────────────────────┘         └─────────────┘

When OpenRouter rate-limits you, the next request goes to Groq. When Groq's daily token cap hits, the next goes to HuggingFace. Your agent never sees a 429.

Why this exists

You can already get a free tier from OpenRouter. And NVIDIA. And Groq. And Cloudflare Workers AI. And HuggingFace. They all have different limits, different free-detection rules, different ways of saying "you're done for today."

So you sign up for all of them and now you've got five API keys, five SDKs, and an agent that only knows about one. FreeRide is the small thing that sits between them and pretends to be one OpenAI endpoint.

  • Local-first. The gateway runs on your machine. Prompts and completions never touch a FreeRide server.
  • BYO keys. Bring your own free-tier keys. FreeRide doesn't issue any.
  • Free-only. No paid fallback. No upsell. If every provider is exhausted, the request fails — better that than a surprise bill.

Install

curl -sSL https://free-ride.xyz/install.sh | sh

The installer bootstraps uv if missing, then uv tool installs freeride-gateway. Binary lands at ~/.local/bin/freeride. Same shape as the bun.sh and astral.sh installers.

Or install manually
# uv (what the installer does)
uv tool install --prerelease=allow freeride-gateway

# pipx
pipx install --pip-args=--pre freeride-gateway

# pip + venv (the venv only — re-activate per shell)
python3 -m venv .venv && source .venv/bin/activate
pip install --pre freeride-gateway

# from source
git clone https://github.com/Shaivpidadi/FreeRideV3 && cd FreeRideV3
pip install -e .

PyPI distribution: freeride-gateway. CLI: freeride. Python ≥ 3.10.

Get keys (any one is enough; more = better failover)

Provider Where Env var
OpenRouter https://openrouter.ai/keys OPENROUTER_API_KEY
Groq https://console.groq.com/keys GROQ_API_KEY
NVIDIA NIM https://build.nvidia.com NVIDIA_API_KEY
Cloudflare Workers AI https://dash.cloudflare.com/profile/api-tokens CLOUDFLARE_API_TOKEN + CLOUDFLARE_ACCOUNT_ID
HuggingFace https://huggingface.co/settings/tokens HF_TOKEN

Set whichever you have, then freeride serve. The gateway picks them up and rotates between them.

Wire your agent

The fastest way is a binder:

freeride bind aider       # writes ~/.aider.conf.yml
freeride bind continue    # writes ~/.continue/config.yaml
freeride bind hermes      # writes ~/.hermes/config.yaml
freeride bind openclaw    # writes ~/.openclaw/openclaw.json

Or set the OpenAI vars yourself:

export OPENAI_API_BASE=http://localhost:11343/v1
export OPENAI_API_KEY=any

Anything OpenAI-shaped works. Tested with the openai-python SDK, Aider, Continue, Hermes, OpenClaw.

Multi-key rotation

Got several free keys for the same provider? Pass them as a JSON array:

export OPENROUTER_API_KEY='["sk-or-v1-key1","sk-or-v1-key2","sk-or-v1-key3"]'

When key 1 hits 429 it goes on cooldown for 120s; key 2 takes the next request. Cooldowns persist across restarts (~/.freeride/cooldown.json).

How failover works

Per request, FreeRide walks (provider, key) pairs in order:

  • RATE_LIMIT or AUTH → mark this key cooling, try the next key.
  • MODEL_NOT_FOUND → skip this provider, try the next provider.
  • Anything 5xx-ish → next pair.
  • First successful response → ship it; stamp X-FreeRide-Provider header (or _freeride_provider field on JSON) so you can tell who actually served it.

Streaming uses buffer-first-chunk failover: hold the first SSE event until upstream confirms the stream is real. If it fails before the first chunk, retry. After the first chunk has shipped, mid-stream errors propagate (rare; documented).

Telemetry

On by default. Hourly POST to https://telemetry.free-ride.xyz/v1/beacon:

{
  "installation_id": "random-uuid-v4",
  "version": "0.3.0",
  "os": "darwin",
  "tokens_served": 412034,
  "request_count": 187,
  "providers_active": ["openrouter", "groq"],
  "uptime_hours": 8
}

Prompts, completions, model IDs, API keys, hostnames, IPs — never sent. The Worker doesn't log cf-connecting-ip. The first time you run any freeride command a banner prints the exact payload.

freeride telemetry off    # turn it off
freeride telemetry        # show what would be sent

Commands

freeride serve                  start the gateway
freeride bind <agent>           write gateway URL into agent config
freeride telemetry [on|off]     manage telemetry
freeride list                   list available free models
freeride status                 show OpenClaw config + cache age (v2)
freeride auto                   auto-configure OpenClaw (v2)
freeride rotate                 swap primary if it fails (v2)
freeride-watcher                background daemon that rotates on failure

The v2 commands keep working for existing OpenClaw users.

Providers

Provider Status Notes
OpenRouter shipped full surface — chat, streaming, tools, vision, structured outputs
NVIDIA NIM shipped curated free-model allowlist; NVIDIA_NIM_FREE_MODELS_OVERRIDE to expand
Groq shipped hardcoded allowlist (Llama 3.x, Gemma 2, Mixtral, DeepSeek-R1-distill); GROQ_FREE_MODELS_OVERRIDE to expand
Cloudflare Workers AI shipped curated allowlist of cheap-per-neuron chat models; needs CLOUDFLARE_ACCOUNT_ID
HuggingFace Inference shipped full HF router catalog; budget governs access ($0.10/mo Free, $2/mo PRO)

Adding a sixth: implement freeride.core.provider.Provider (api_version=1) in freeride/providers/<name>.py, register it in the conformance suite, done. See CONTRIBUTING.md.

Agents

Agent freeride bind Hot reload
OpenClaw yes needs restart
Aider yes (--scope home/cwd/git) needs restart
Continue yes yes
Hermes (NousResearch/hermes-agent) yes needs restart

Or anything else: OPENAI_API_BASE=http://localhost:11343/v1 + OPENAI_API_KEY=any.

Docs

License

MIT.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

freeride_gateway-0.3.0a7.tar.gz (133.3 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

freeride_gateway-0.3.0a7-py3-none-any.whl (75.4 kB view details)

Uploaded Python 3

File details

Details for the file freeride_gateway-0.3.0a7.tar.gz.

File metadata

  • Download URL: freeride_gateway-0.3.0a7.tar.gz
  • Upload date:
  • Size: 133.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for freeride_gateway-0.3.0a7.tar.gz
Algorithm Hash digest
SHA256 fab2a0c4f99abf2b215e76d7bb458143006547f3ebf51c336d2f275fa8edeaa1
MD5 3ab03861f369ba067ff24496776818f7
BLAKE2b-256 851f213f896d41ab40cf92bf5fb5a6efbc8795ef6cc3f714cd00a18e6e91826b

See more details on using hashes here.

Provenance

The following attestation bundles were made for freeride_gateway-0.3.0a7.tar.gz:

Publisher: release.yml on Shaivpidadi/FreeRideV3

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file freeride_gateway-0.3.0a7-py3-none-any.whl.

File metadata

File hashes

Hashes for freeride_gateway-0.3.0a7-py3-none-any.whl
Algorithm Hash digest
SHA256 96c52f642881af87c05574c52b2b4418b59bf26456d4921441857cbd96f5ef6a
MD5 56df9851eb2dfa35b15b8488fef46d61
BLAKE2b-256 52b43d2301ede772349e7a495c9f3b0f080bbc62352128040a4a5c3a0920fc4a

See more details on using hashes here.

Provenance

The following attestation bundles were made for freeride_gateway-0.3.0a7-py3-none-any.whl:

Publisher: release.yml on Shaivpidadi/FreeRideV3

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page