Skip to main content

Cut your AI token costs by 40-60%. Drop-in proxy for OpenAI, LangChain, LlamaIndex, CrewAI.

Project description

AgentReady — Python SDK

Cut AI token costs by 40-60%. Drop-in proxy for OpenAI, LangChain, LlamaIndex, CrewAI.

PyPI Python License: MIT

Install & Setup (30 seconds)

pip install agentready-sdk
agentready init

This opens your browser → create a free account → API key is saved to .env automatically.

Quick Start — Drop-in Proxy (Recommended)

Just swap your base_url. Zero code changes to your existing OpenAI calls:

from openai import OpenAI

client = OpenAI(
    base_url="https://agentready.cloud/v1",   # ← only change needed
    api_key="ak_...",                          # your AgentReady key
    default_headers={
        "X-Upstream-API-Key": "sk-...",        # your OpenAI key
    },
)

# Everything works exactly like before — but 40-60% cheaper
response = client.chat.completions.create(
    model="gpt-4o",
    messages=[{"role": "user", "content": your_long_prompt}],
)

One-liner Helper

import agentready

client = agentready.openai("ak_...", upstream_key="sk-...")
response = client.chat.completions.create(
    model="gpt-4o",
    messages=[{"role": "user", "content": "Hello!"}],
)

Async Client

client = agentready.create_client("ak_...", upstream_key="sk-...", async_client=True)
response = await client.chat.completions.create(...)

Method 2 — Monkey-Patch

Patch all OpenAI/Anthropic calls globally with two lines:

from agentready import patch_openai
patch_openai(api_key="ak_...")

# All existing OpenAI code is now compressed automatically
from openai import OpenAI
client = OpenAI()
response = client.chat.completions.create(
    model="gpt-4o",
    messages=[{"role": "user", "content": your_long_prompt}],
)

Or patch everything at once:

import agentready
agentready.api_key = "ak_..."
agentready.auto()  # patches OpenAI + Anthropic

Method 3 — Manual Compression

For fine-grained control:

import agentready
agentready.api_key = "ak_..."

result = agentready.compress("Your long prompt text here...")
print(result.text)              # compressed text
print(result.tokens_saved)      # 1,247
print(result.reduction_percent) # 52.3
print(result.savings_usd)       # 0.0374

Framework Integrations

LangChain

from agentready.integrations.langchain import TokenCutCallbackHandler
from langchain_openai import ChatOpenAI

handler = TokenCutCallbackHandler(api_key="ak_...")
llm = ChatOpenAI(model="gpt-4o", callbacks=[handler])
response = llm.invoke("Your very long prompt here...")

LlamaIndex

from agentready.integrations.llamaindex import TokenCutPostprocessor

postprocessor = TokenCutPostprocessor(api_key="ak_...")
query_engine = index.as_query_engine(
    node_postprocessors=[postprocessor]
)

CrewAI

from agentready.integrations.crewai import create_crewai_llm
from crewai import Agent, Task, Crew

llm = create_crewai_llm(
    agentready_key="ak_...",
    upstream_key="sk-...",
    model="gpt-4o",
)

agent = Agent(
    role="Researcher",
    goal="Research AI trends",
    backstory="Expert AI researcher.",
    llm=llm,
)

How It Works

AgentReady's proxy sits between your code and OpenAI. Every request is:

  1. Compressed — redundant phrasing removed, verbose text condensed
  2. Forwarded — sent to OpenAI with your upstream key
  3. Returned — response comes back unchanged

Code blocks, URLs, numbers, and key terms are always preserved.

Configuration

# Proxy mode — compression level via header
client = agentready.openai(
    "ak_...",
    upstream_key="sk-...",
    compression_level="aggressive",  # "light", "standard", "aggressive"
)

# Patch mode — configuration via arguments
agentready.auto(
    level="medium",
    preserve_code=True,
    min_tokens=100,
)

Pricing

Beta — Free unlimited usage. After beta: pay-per-token, ~60% less than direct API costs.

Get your API key at agentready.cloud

License

MIT — AgentReady

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

agentready_sdk-0.4.0.tar.gz (14.7 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

agentready_sdk-0.4.0-py3-none-any.whl (17.1 kB view details)

Uploaded Python 3

File details

Details for the file agentready_sdk-0.4.0.tar.gz.

File metadata

  • Download URL: agentready_sdk-0.4.0.tar.gz
  • Upload date:
  • Size: 14.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.14.2

File hashes

Hashes for agentready_sdk-0.4.0.tar.gz
Algorithm Hash digest
SHA256 12d4e09d673f4540bdf02d37bc71f63a0418ba9d253e9327521e5c4c061d9cca
MD5 d213103ce07daf4067edfc9f05f241c4
BLAKE2b-256 d6dee0e3efe2c08a6a186f6945099c928ba74289ba1fc17c0b96982f53e3d306

See more details on using hashes here.

File details

Details for the file agentready_sdk-0.4.0-py3-none-any.whl.

File metadata

  • Download URL: agentready_sdk-0.4.0-py3-none-any.whl
  • Upload date:
  • Size: 17.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.14.2

File hashes

Hashes for agentready_sdk-0.4.0-py3-none-any.whl
Algorithm Hash digest
SHA256 6412be9687bb4d9fb57ae1458d860d476f9de9faa8ec5450fd48e47794d1a116
MD5 d22a9f7c61671c04acb191a980f3575d
BLAKE2b-256 1d24c6e60feaa6d2b7e5b31b81cd1f64201ee66a68944367a047c6b232c7c39a

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page