Easy-to-use LLM API from a state-of-the-art provider and comparison
Project description
api4all
Easy-to-use LLM API from a state-of-the-art provider and comparison.
Installation
- Install the package
pip3 install api4all
- Create and activate a virtual environment (optional but recommended)
- Unix / macOS
python3 -m venv venv
source venv/bin/activate
- Windows
python3 -m venv venv
.\venv\Scripts\activate
Quick Start
- Wrap the API key in a
.env
file of the provider you want to test.
TOGETHER_API_KEY=xxx
OPENAI_API_KEY=xxx
MISTRAL_API_KEY=xxx
ANTHROPIC_API_KEY=xxx
- Run the code
from api4all import EngineFactory
messages = [
{"role": "system",
"content": "You are a helpful assistent for the my Calculus class."},
{"role": "user",
"content": "What is the current status of the economy?"}
]
engine = EngineFactory.create_engine(provider="together",
model="google/gemma-7b-it",
messages=messages,
temperature=0.5,
max_tokens=256,
)
response = engine.generate_response()
print(response)
- Check the log file for the response and the cost of the request.
Request ID - fa8cebd0-265a-44b2-95d7-6ff1588d2c87
create at: 2024-03-15 16:38:18,129
INFO - SUCCESS
Response:
I am not able to provide information about the current status of the economy, as I do not have access to real-time information. Therefore, I recommend checking a reliable source for the latest economic news and data.
Cost: $0.0000154 # Cost of this provider for this request
Provider: together # Provider used for this request
Execution-time: Execution time not provided by the provider
Actual-time: 0.9448428153991699 # Actual time taken by the request
Input-token: 33 # Number of tokens used for the input
Output-token: 44 # Number of tokens used for the output
Providers and Pricing
Providers
Provider | Free Credit | Rate Limit | API Key name | Provider string name |
---|---|---|---|---|
Groq | Unlimited | 30 Requests / Minute | GROQ_API_KEY | "groq" |
Anyscale | $10 | 30 Requests / Second | ANYSCALE_API_KEY | "anyscale" |
Together AI | $25 | 1 Requests / Second | TOGETHER_API_KEY | "together" |
Replicate | Free to try | 50 Requests / Second | REPLICATE_API_KEY | "replicate" |
Fireworks | $1 | 600 Requests / Minute | FIREWORKS_API_KEY | "fireworks" |
Deepinfra | Free to try | 200 Concurrent request | DEEPINFRA_API_KEY | "deepinfra" |
Google AI (Vertex AI) | Unlimited | 60 Requests / Minute | GOOGLE_API_KEY | "google" |
OpenAI | ✕ | 60 Requests / Minute | OPENAI_API_KEY | "openai" |
Mistral AI | Free to try | 5 Requests / Second | MISTRAL_API_KEY | "mistral" |
Anthropic | Free to try | 5 Requests / Minute | ANTHROPIC_API_KEY | "anthropic" |
- Free to try: Free to try, no credit card required but limited to a certain number of tokens.
Open-source models
1. Mixtral-8x7b-Instruct-v0.1
Mixtral 8x7B is a high-performing sparse mixture of experts model (SMoE) with open weights, licensed under Apache 2.0. It surpasses Llama 2 70B in most benchmarks, offering 6x faster inference. It's the best open-weight model considering cost/performance trade-offs, matching or exceeding GPT3.5 on most standard benchmarks.
- Developer: Mistral AI
- Context Length: 32,768
Provider | Input Pricing ($/1M Tokens) | Output Pricing ($/1M Tokens) |
---|---|---|
Groq | $0 | $0 |
Anyscale | $0.5 | $0.5 |
Together AI | $0.6 | $0.6 |
Replicate | $0.3 | $1 |
Fireworks | $0.5 | $0.5 |
Deepinfra | $0.27 | $0.27 |
2. Gemma 7B it
- Developer: Google AI
- Context Length: 8,192
Provider | Input Pricing ($/1M Tokens) | Output Pricing ($/1M Tokens) |
---|---|---|
Groq | $0 | $0 |
Anyscale | $0.15 | $0.15 |
Together AI | $0.2 | $0.2 |
Replicate | ✕ | ✕ |
Fireworks | $0.2 | $0.2 |
Deepinfra | ✕ | ✕ |
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
api4all-0.2.8.tar.gz
(14.1 kB
view details)
Built Distribution
api4all-0.2.8-py3-none-any.whl
(14.5 kB
view details)
File details
Details for the file api4all-0.2.8.tar.gz
.
File metadata
- Download URL: api4all-0.2.8.tar.gz
- Upload date:
- Size: 14.1 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.9.13
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 1224292123b8d8e224f1a2d875fff127eaf998e4b5de14de8c8df0c32a1cf600 |
|
MD5 | 06430ae272880787929c43a2d0bee17e |
|
BLAKE2b-256 | 56b2730f1cb0cdfef1e07f975d8e65105fbbc14a1cba70149e24d1bd6c4efca1 |
File details
Details for the file api4all-0.2.8-py3-none-any.whl
.
File metadata
- Download URL: api4all-0.2.8-py3-none-any.whl
- Upload date:
- Size: 14.5 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.9.13
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 9025a944daca2cb90693eed8c04c6d252be8f19b5075a5045777f4f7723506a3 |
|
MD5 | f45f643c8f092ff466e6a6786516ed16 |
|
BLAKE2b-256 | 9190067d63a78a8964f9cb9516337e780a61210bf8c028514ddbd2ce0ed493d7 |