Reduce LLM costs by 90% - AI recommendations with NO API keys needed!
Project description
๐ LLMOptimize - AI Cost Optimization Made Simple
Reduce your LLM API costs by 90% with zero configuration. No API keys required for reports!
โจ What Makes LLMOptimize Different?
๐ No API Keys Needed!
Unlike other tools, you don't need OpenAI, Anthropic, or any API keys to get cost reports and smart recommendations. Our server does the AI analysis using our keys!
๐ฏ Smart, Context-Aware Recommendations
- Track embeddings? Get embedding alternatives (not chat models!)
- Track chat? Get chat alternatives
- Track reasoning models? Get reasoning alternatives
- ML-powered suggestions that learn from your patterns
๐ Beautiful Interactive Reports
Gorgeous terminal UI with animations, color-coded insights, and actionable recommendations.
๐ฌ Quick Demo
import llmoptimize
import openai
# Use OpenAI normally - LLMOptimize tracks automatically!
client = openai.OpenAI()
response = client.chat.completions.create(
model="gpt-4",
messages=[{"role": "user", "content": "Hello!"}]
)
# See your beautiful cost report (NO API KEY NEEDED!)
llmoptimize.report()
Output:
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ โ
โ ๐ L L M O P T I M I Z E R E P O R T ๐ โ
โ โ
โ Recommendations from OUR AI (No Key Needed!) โ
โ โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
๐ YOUR USAGE SUMMARY
๐ Total API Calls Tracked
15
๐ฐ Total Cost
$0.1047
๐ Potential Savings
$0.0938
That's 90% you could save!
๐ MODEL USAGE
gpt-4: 10 calls
text-embedding-3-large: 5 calls
๐ก SMART RECOMMENDATIONS
(From OUR AI - no API key needed!)
โญโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโฎ
โ #1 Recommendation โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค
โ ๐ฏ gpt-4o-mini โ
โ ๐ฐ Save 99% โ
โ โ
โ ๐ก Used gpt-4 10x - gpt-4o-mini is 99% cheaper โ
โฐโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโฏ
โญโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโฎ
โ #2 Recommendation โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค
โ ๐ฏ text-embedding-3-small โ
โ ๐ฐ Save 85% โ
โ โ
โ ๐ก Used text-embedding-3-large 5x - smaller model โ
โ is 85% cheaper with 95%+ quality โ
โฐโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโฏ
โจ All powered by OUR server intelligence!
๐ Installation
pip install llmoptimize
That's it! No configuration needed. Just import and use.
๐ Features
๐ฏ Auto-Tracking
- Automatically tracks OpenAI, Anthropic, Groq, and LangChain
- Zero configuration required
- Works with existing code - no changes needed
- Supports: Chat, Embeddings, Images, Audio
๐ค Smart Recommendations
- ML-powered model suggestions
- Context-aware (embeddings โ embeddings, chat โ chat)
- Learns from your usage patterns
- Based on 50+ AI models across all providers
๐ Beautiful Dashboards
- Interactive animated terminal UI
- Real-time cost tracking
- Usage analytics by model
- Weekly usage timelines
๐ Privacy-First Design
- No prompts sent to our server
- Only metadata tracked (model, tokens)
- Anonymous session IDs
- GDPR compliant
๐ก No API Keys Required
- Reports work without ANY API keys
- AI recommendations from OUR server
- You keep your keys, we provide the intelligence
๐ฏ Supported Providers & Models
OpenAI
- โ Chat: GPT-4, GPT-4o, GPT-4o-mini, GPT-3.5-turbo, o1, o1-mini, o3-mini
- โ Embeddings: text-embedding-3-large, text-embedding-3-small, ada-002
- โ Images: DALL-E 3, DALL-E 2
- โ Audio: Whisper
Anthropic
- โ Claude 3.5: Opus, Sonnet, Haiku
- โ Claude 3: Opus, Sonnet, Haiku
- โ Gemini 2.0: Pro, Flash, Flash-Lite
- โ Gemini 1.5: Pro, Flash, Flash-8B
Groq
- โ Llama 3.3 70B, Llama 3.1 70B/8B
- โ Mixtral 8x7B
- โ Gemma 2 9B
Mistral
- โ Large, Medium, Small, Nemo, Codestral
- โ Open Mixtral, Open Mistral
Cohere
- โ Command R+, Command R, Command
Total: 50+ models tracked with accurate pricing!
๐ Usage Examples
Example 1: Chat Completion
import llmoptimize
import openai
client = openai.OpenAI()
# Make expensive call
response = client.chat.completions.create(
model="gpt-4", # $0.03 per 1K input tokens
messages=[{"role": "user", "content": "Explain quantum computing"}],
max_tokens=500
)
# See recommendations
llmoptimize.report()
# Suggests: gpt-4o-mini (99% cheaper!) or gpt-3.5-turbo (95% cheaper)
Example 2: Embeddings
import llmoptimize
import openai
client = openai.OpenAI()
# Expensive embeddings
embeddings = client.embeddings.create(
model="text-embedding-3-large", # $0.13 per 1M tokens
input=["Document 1", "Document 2", "Document 3"]
)
llmoptimize.report()
# Suggests: text-embedding-3-small (85% cheaper with 95%+ quality)
Example 3: Anthropic Claude
import llmoptimize
import anthropic
client = anthropic.Anthropic()
# Expensive Claude Opus
response = client.messages.create(
model="claude-3-opus-20240229", # $15 per 1M input tokens
max_tokens=1000,
messages=[{"role": "user", "content": "Write a poem"}]
)
llmoptimize.report()
# Suggests: claude-3-5-haiku (95% cheaper) or claude-3-haiku (98% cheaper)
Example 4: Mixed Usage
import llmoptimize
import openai
client = openai.OpenAI()
# Multiple different calls
client.chat.completions.create(model="gpt-4", messages=[...])
client.chat.completions.create(model="gpt-3.5-turbo", messages=[...])
client.embeddings.create(model="text-embedding-3-large", input=[...])
llmoptimize.report()
# Gets CONTEXT-AWARE recommendations for each model type!
Example 5: Simple Text Report
import llmoptimize
# For scripts/logs - simple text output
llmoptimize.report(interactive=False)
Output:
๐ LLMOptimize Report
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
Calls: 15
Cost: $0.1047
Savings: $0.0938 (90%)
Models:
gpt-4: 10
text-embedding-3-large: 5
Recommendations (from OUR AI):
1. gpt-4o-mini - 99%
2. text-embedding-3-small - 85%
No API key needed!
Example 6: Manual Tracking
import llmoptimize
# For custom APIs or non-supported providers
llmoptimize.track(
model="custom-model-v1",
prompt_tokens=1000,
completion_tokens=500,
provider="custom"
)
llmoptimize.report()
๐จ Advanced Features
1. LangChain Integration
import llmoptimize
from langchain_openai import ChatOpenAI
# LangChain auto-tracked!
llm = ChatOpenAI(model="gpt-4")
result = llm.invoke("Hello!")
llmoptimize.report()
# Automatically tracks LangChain calls
2. Agent Workflow Tracking
import llmoptimize
# Track multi-step agent workflows
# Loop detection, context window monitoring
# Step-by-step cost breakdown
llmoptimize.report()
# Shows per-step costs and optimization opportunities
3. Real-Time Monitoring
import llmoptimize
# In production scripts
while True:
# Your AI calls...
# Periodic cost check
if call_count % 100 == 0:
llmoptimize.report(interactive=False)
๐ Real Results
Case Study 1: SaaS Startup
"We reduced our OpenAI bill from $4,200/month to $380/month using LLMOptimize recommendations. The ROI was immediate."
โ AI-powered SaaS, 50K users
Changes made:
- GPT-4 โ GPT-4o-mini for 80% of tasks: 99% savings
- text-embedding-3-large โ text-embedding-3-small: 85% savings
- GPT-4 โ Claude-3-Haiku for simple Q&A: 98% savings
Case Study 2: AI Research Team
"The interactive dashboard makes it so easy to spot optimization opportunities. Saved us 92% on GPT-4 costs without quality loss."
โ University AI Lab
Changes made:
- GPT-4 โ GPT-3.5-turbo for data classification: 95% savings
- Claude-3-Opus โ Claude-3-Sonnet for analysis: 80% savings
- Eliminated redundant embedding calls: 100% savings
Case Study 3: E-commerce Platform
"We were using GPT-4 for EVERYTHING. LLMOptimize showed us where GPT-3.5-turbo or even Groq's free tier would work. $2,800/month โ $290/month."
โ E-commerce AI chatbot, 100K conversations/month
๐๏ธ How It Works
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ 1. YOU: Import llmoptimize โ
โ โ Auto-patches OpenAI/Anthropic/Groq SDKs โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ 2. YOU: Make API calls normally โ
โ โ client.chat.completions.create(...) โ
โ โ client.embeddings.create(...) โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ 3. LLMOptimize: Tracks metadata โ
โ โ Model name, token counts โ
โ โ Sends to OUR server (not your API!) โ
โ โ NO prompts sent (privacy-first!) โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ 4. OUR Server: Analyzes with OUR AI โ
โ โ Uses ML + 50+ model pricing data โ
โ โ Context-aware recommendations โ
โ โ Learning from usage patterns โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ 5. YOU: See beautiful report โ
โ โ llmoptimize.report() โ
โ โ NO API KEY NEEDED! โ
โ โ Smart, actionable recommendations โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
๐ Privacy & Security
What We Track:
- โ Model name (e.g., "gpt-4")
- โ Token counts (input/output)
- โ Anonymous session ID
- โ Provider (OpenAI/Anthropic/etc)
What We DON'T Track:
- โ Your prompts or AI responses
- โ Personal information
- โ API keys
- โ IP addresses
- โ User identifiers
Data Storage:
- ๐ Anonymous sessions only
- ๐ No PII collected
- ๐ GDPR compliant
- ๐ Can run self-hosted if needed
๐ฏ Why LLMOptimize?
| Feature | LLMOptimize | Competitors |
|---|---|---|
| No API Keys for Reports | โ Our AI, Our Keys | โ Need your keys |
| Context-Aware Recommendations | โ Embeddings โ Embeddings | โ Generic suggestions |
| Auto-Tracking | โ Zero config | โ ๏ธ Manual setup |
| Beautiful UI | โ Animated terminal | โ Plain text |
| 50+ Models | โ All providers | โ ๏ธ OpenAI only |
| Privacy-First | โ No prompts sent | โ ๏ธ Varies |
| ML Learning | โ Improves over time | โ Static rules |
| Free Tier | โ Unlimited tracking | โ ๏ธ Limited |
๐ ๏ธ Configuration
Environment Variables (Optional)
# Custom server URL (for self-hosted)
export LLMOPTIMIZE_SERVER_URL=https://your-server.com
# Session ID (auto-generated if not set)
export LLMOPTIMIZE_SESSION_ID=your-session-id
Programmatic Config
import llmoptimize
# Use custom server
llmoptimize.SERVER_URL = "https://your-server.com"
# Custom session
llmoptimize.SESSION_ID = "my-session"
๐ฆ Installation Options
Basic (Default)
pip install llmoptimize
With All Provider SDKs
pip install llmoptimize[full]
# Includes: anthropic, groq, langchain
Development Tools
pip install llmoptimize[dev]
# Includes: pytest, black, mypy
๐ค Contributing
This is a proprietary project. For bugs and feature requests, please open an issue on GitHub.
๐ License
Proprietary - All Rights Reserved
๐ Links
- Homepage: https://aioptimize.up.railway.app
- GitHub: https://github.com/hackrudra1234/llmoptimize
- PyPI: https://pypi.org/project/llmoptimize
- Documentation: Coming soon!
๐ฌ Support
- Issues: GitHub Issues
- Email: hackrudra@gmail.com
- Twitter: Coming soon!
๐ Get Started Now!
# Install
pip install llmoptimize
# Use
import llmoptimize
# ... your AI code ...
llmoptimize.report()
Save 90% on AI costs. Beautiful reports. Zero configuration. No API keys needed! ๐
Made with โค๏ธ by the LLMOptimize Team
Star us on GitHub if this saves you money! โญ
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file llmoptimize-3.2.0.tar.gz.
File metadata
- Download URL: llmoptimize-3.2.0.tar.gz
- Upload date:
- Size: 11.9 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
558ab4c7548deb7b71e69c64691282993aaf3e16fc0c20ebeebc0cb52389a06a
|
|
| MD5 |
66a9bd05397658351ce6333d74968482
|
|
| BLAKE2b-256 |
422e92ab1ce7231422ac8c16f2754275e30dcab9faf9f751a016717821f18a45
|
File details
Details for the file llmoptimize-3.2.0-py3-none-any.whl.
File metadata
- Download URL: llmoptimize-3.2.0-py3-none-any.whl
- Upload date:
- Size: 10.0 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
418db9b20c1f9fd4021c4abaad3b678d97014b9830dcfa2bebd8b5a05258ca3e
|
|
| MD5 |
8296e6a04766a418a296995bb733978c
|
|
| BLAKE2b-256 |
489a35b60b97af21fdbb498a3335de4cf07dc8e11a694d28c41171801eb5367c
|