Skip to main content

Cut your AI token costs by 40-60%. Drop-in proxy for OpenAI, LangChain, LlamaIndex, CrewAI.

Project description

AgentReady — Python SDK

Cut AI token costs by 40-60%. Drop-in proxy for OpenAI, LangChain, LlamaIndex, CrewAI.

PyPI Python License: MIT

Install

pip install agentready

Quick Start — Drop-in Proxy (Recommended)

Just swap your base_url. Zero code changes to your existing OpenAI calls:

from openai import OpenAI

client = OpenAI(
    base_url="https://agentready.cloud/v1",   # ← only change needed
    api_key="ak_...",                          # your AgentReady key
    default_headers={
        "X-Upstream-API-Key": "sk-...",        # your OpenAI key
    },
)

# Everything works exactly like before — but 40-60% cheaper
response = client.chat.completions.create(
    model="gpt-4o",
    messages=[{"role": "user", "content": your_long_prompt}],
)

One-liner Helper

import agentready

client = agentready.openai("ak_...", upstream_key="sk-...")
response = client.chat.completions.create(
    model="gpt-4o",
    messages=[{"role": "user", "content": "Hello!"}],
)

Async Client

client = agentready.create_client("ak_...", upstream_key="sk-...", async_client=True)
response = await client.chat.completions.create(...)

Method 2 — Monkey-Patch

Patch all OpenAI/Anthropic calls globally with two lines:

from agentready import patch_openai
patch_openai(api_key="ak_...")

# All existing OpenAI code is now compressed automatically
from openai import OpenAI
client = OpenAI()
response = client.chat.completions.create(
    model="gpt-4o",
    messages=[{"role": "user", "content": your_long_prompt}],
)

Or patch everything at once:

import agentready
agentready.api_key = "ak_..."
agentready.auto()  # patches OpenAI + Anthropic

Method 3 — Manual Compression

For fine-grained control:

import agentready
agentready.api_key = "ak_..."

result = agentready.compress("Your long prompt text here...")
print(result.text)              # compressed text
print(result.tokens_saved)      # 1,247
print(result.reduction_percent) # 52.3
print(result.savings_usd)       # 0.0374

Framework Integrations

LangChain

from agentready.integrations.langchain import TokenCutCallbackHandler
from langchain_openai import ChatOpenAI

handler = TokenCutCallbackHandler(api_key="ak_...")
llm = ChatOpenAI(model="gpt-4o", callbacks=[handler])
response = llm.invoke("Your very long prompt here...")

LlamaIndex

from agentready.integrations.llamaindex import TokenCutPostprocessor

postprocessor = TokenCutPostprocessor(api_key="ak_...")
query_engine = index.as_query_engine(
    node_postprocessors=[postprocessor]
)

CrewAI

from agentready.integrations.crewai import create_crewai_llm
from crewai import Agent, Task, Crew

llm = create_crewai_llm(
    agentready_key="ak_...",
    upstream_key="sk-...",
    model="gpt-4o",
)

agent = Agent(
    role="Researcher",
    goal="Research AI trends",
    backstory="Expert AI researcher.",
    llm=llm,
)

How It Works

AgentReady's proxy sits between your code and OpenAI. Every request is:

  1. Compressed — redundant phrasing removed, verbose text condensed
  2. Forwarded — sent to OpenAI with your upstream key
  3. Returned — response comes back unchanged

Code blocks, URLs, numbers, and key terms are always preserved.

Configuration

# Proxy mode — compression level via header
client = agentready.openai(
    "ak_...",
    upstream_key="sk-...",
    compression_level="aggressive",  # "light", "standard", "aggressive"
)

# Patch mode — configuration via arguments
agentready.auto(
    level="medium",
    preserve_code=True,
    min_tokens=100,
)

Pricing

Beta — Free unlimited usage. After beta: pay-per-token, ~60% less than direct API costs.

Get your API key at agentready.cloud

License

MIT — AgentReady

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

agentready_sdk-0.3.0.tar.gz (11.9 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

agentready_sdk-0.3.0-py3-none-any.whl (13.9 kB view details)

Uploaded Python 3

File details

Details for the file agentready_sdk-0.3.0.tar.gz.

File metadata

  • Download URL: agentready_sdk-0.3.0.tar.gz
  • Upload date:
  • Size: 11.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.14.2

File hashes

Hashes for agentready_sdk-0.3.0.tar.gz
Algorithm Hash digest
SHA256 de9686878b02f207d012cf4485b190344b7d66540de4d46ec4911cafa44fca1e
MD5 1b43ea0f9e6f40d3370a8196e6dbdc61
BLAKE2b-256 47a498a887d59beb9831808aba305549df53aa4e920b9b5c70727d66b9eac60f

See more details on using hashes here.

File details

Details for the file agentready_sdk-0.3.0-py3-none-any.whl.

File metadata

  • Download URL: agentready_sdk-0.3.0-py3-none-any.whl
  • Upload date:
  • Size: 13.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.14.2

File hashes

Hashes for agentready_sdk-0.3.0-py3-none-any.whl
Algorithm Hash digest
SHA256 856c83d3b3f81432f5012bf79f6805b41f1f97bc19fc5ffe2b6f0e2cb9598699
MD5 48f8f54cb4890a60171b589e14afbf85
BLAKE2b-256 388da04006a927ae51f8604607c4d4072c84e2660800e385755fd62b0b082d04

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page