Cut your AI token costs by 40-60%. Drop-in proxy for OpenAI, LangChain, LlamaIndex, CrewAI.
Project description
AgentReady — Python SDK
Cut AI token costs by 40-60%. Drop-in proxy for OpenAI, LangChain, LlamaIndex, CrewAI.
Install
pip install agentready
Quick Start — Drop-in Proxy (Recommended)
Just swap your base_url. Zero code changes to your existing OpenAI calls:
from openai import OpenAI
client = OpenAI(
base_url="https://agentready.cloud/v1", # ← only change needed
api_key="ak_...", # your AgentReady key
default_headers={
"X-Upstream-API-Key": "sk-...", # your OpenAI key
},
)
# Everything works exactly like before — but 40-60% cheaper
response = client.chat.completions.create(
model="gpt-4o",
messages=[{"role": "user", "content": your_long_prompt}],
)
One-liner Helper
import agentready
client = agentready.openai("ak_...", upstream_key="sk-...")
response = client.chat.completions.create(
model="gpt-4o",
messages=[{"role": "user", "content": "Hello!"}],
)
Async Client
client = agentready.create_client("ak_...", upstream_key="sk-...", async_client=True)
response = await client.chat.completions.create(...)
Method 2 — Monkey-Patch
Patch all OpenAI/Anthropic calls globally with two lines:
from agentready import patch_openai
patch_openai(api_key="ak_...")
# All existing OpenAI code is now compressed automatically
from openai import OpenAI
client = OpenAI()
response = client.chat.completions.create(
model="gpt-4o",
messages=[{"role": "user", "content": your_long_prompt}],
)
Or patch everything at once:
import agentready
agentready.api_key = "ak_..."
agentready.auto() # patches OpenAI + Anthropic
Method 3 — Manual Compression
For fine-grained control:
import agentready
agentready.api_key = "ak_..."
result = agentready.compress("Your long prompt text here...")
print(result.text) # compressed text
print(result.tokens_saved) # 1,247
print(result.reduction_percent) # 52.3
print(result.savings_usd) # 0.0374
Framework Integrations
LangChain
from agentready.integrations.langchain import TokenCutCallbackHandler
from langchain_openai import ChatOpenAI
handler = TokenCutCallbackHandler(api_key="ak_...")
llm = ChatOpenAI(model="gpt-4o", callbacks=[handler])
response = llm.invoke("Your very long prompt here...")
LlamaIndex
from agentready.integrations.llamaindex import TokenCutPostprocessor
postprocessor = TokenCutPostprocessor(api_key="ak_...")
query_engine = index.as_query_engine(
node_postprocessors=[postprocessor]
)
CrewAI
from agentready.integrations.crewai import create_crewai_llm
from crewai import Agent, Task, Crew
llm = create_crewai_llm(
agentready_key="ak_...",
upstream_key="sk-...",
model="gpt-4o",
)
agent = Agent(
role="Researcher",
goal="Research AI trends",
backstory="Expert AI researcher.",
llm=llm,
)
How It Works
AgentReady's proxy sits between your code and OpenAI. Every request is:
- Compressed — redundant phrasing removed, verbose text condensed
- Forwarded — sent to OpenAI with your upstream key
- Returned — response comes back unchanged
Code blocks, URLs, numbers, and key terms are always preserved.
Configuration
# Proxy mode — compression level via header
client = agentready.openai(
"ak_...",
upstream_key="sk-...",
compression_level="aggressive", # "light", "standard", "aggressive"
)
# Patch mode — configuration via arguments
agentready.auto(
level="medium",
preserve_code=True,
min_tokens=100,
)
Pricing
Beta — Free unlimited usage. After beta: pay-per-token, ~60% less than direct API costs.
Get your API key at agentready.cloud
License
MIT — AgentReady
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file agentready_sdk-0.3.0.tar.gz.
File metadata
- Download URL: agentready_sdk-0.3.0.tar.gz
- Upload date:
- Size: 11.9 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.14.2
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
de9686878b02f207d012cf4485b190344b7d66540de4d46ec4911cafa44fca1e
|
|
| MD5 |
1b43ea0f9e6f40d3370a8196e6dbdc61
|
|
| BLAKE2b-256 |
47a498a887d59beb9831808aba305549df53aa4e920b9b5c70727d66b9eac60f
|
File details
Details for the file agentready_sdk-0.3.0-py3-none-any.whl.
File metadata
- Download URL: agentready_sdk-0.3.0-py3-none-any.whl
- Upload date:
- Size: 13.9 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.14.2
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
856c83d3b3f81432f5012bf79f6805b41f1f97bc19fc5ffe2b6f0e2cb9598699
|
|
| MD5 |
48f8f54cb4890a60171b589e14afbf85
|
|
| BLAKE2b-256 |
388da04006a927ae51f8604607c4d4072c84e2660800e385755fd62b0b082d04
|