Skip to main content

Easy-to-use LLM API from a state-of-the-art provider and comparison

Project description

api4all

Easy-to-use LLM API from a state-of-the-art provider and comparison.

Installation

  1. Clone the repository
git clone https://github.com/api4all/api4all.git
  1. Create and activate a virtual environment
  • Unix / macOS
python3 -m venv venv
source venv/bin/activate
  • Windows
python3 -m venv venv
.\venv\Scripts\activate
  1. Install the requirements
pip install -r requirements.txt

Providers and Pricing

Providers

Provider Free Credit Rate Limit API Key name Provider string name
Groq Unlimited 30 Requests / Minute GROQ_API_KEY "groq"
Anyscale $10 30 Requests / Second ANYSCALE_API_KEY "anyscale"
Together AI $25 1 Requests / Second TOGETHER_API_KEY "together"
Replicate Free to try 50 Requests / Second REPLICATE_API_KEY "replicate"
Fireworks $1 600 Requests / Minute FIREWORKS_API_KEY "fireworks"
Deepinfra Free to try 200 Concurrent request DEEPINFRA_API_KEY "deepinfra"
Google AI (Vertex AI) Unlimited 60 Requests / Minute GOOGLE_API_KEY "google"
OpenAI 60 Requests / Minute OPENAI_API_KEY "openai"
Mistral AI 5 Requests / Second MISTRAL_API_KEY "mistral"
Anthropic 5 Requests / Minute ANTHROPIC_API_KEY "anthropic"
  • Free to try: Free to try, no credit card required but limited to a certain number of tokens.

Open-source models

1. Mixtral-8x7b-Instruct-v0.1

Mixtral 8x7B is a high-performing sparse mixture of experts model (SMoE) with open weights, licensed under Apache 2.0. It surpasses Llama 2 70B in most benchmarks, offering 6x faster inference. It's the best open-weight model considering cost/performance trade-offs, matching or exceeding GPT3.5 on most standard benchmarks.

  • Developer: Mistral AI
  • Context Length: 32,768
Provider Input Pricing ($/1M Tokens) Output Pricing ($/1M Tokens)
Groq $0 $0
Anyscale $0.5 $0.5
Together AI $0.6 $0.6
Replicate $0.3 $1
Fireworks $0.5 $0.5
Deepinfra $0.27 $0.27
2. Gemma 7B it
  • Developer: Google AI
  • Context Length: 8,192
Provider Input Pricing ($/1M Tokens) Output Pricing ($/1M Tokens)
Groq $0 $0
Anyscale $0.15 $0.15
Together AI $0.2 $0.2
Replicate
Fireworks $0.2 $0.2
Deepinfra
3. LLaMA2-70b
  • Developer: Meta AI
  • Context Length: 4,096
Provider Input Pricing ($/1M Tokens) Output Pricing ($/1M Tokens)
Groq $0 $0
Anyscale $1.0 $1.0
Together AI $0.9 $0.9
Replicate $0.65 $2.75
Fireworks $0.9 $0.9
Deepinfra $0.7 $0.9
4. Mistral-7B-Instruct-v0.1
  • Developer: Mistral AI
  • Context Length: 4,096
Provider Input Pricing ($/1M Tokens) Output Pricing ($/1M Tokens)
Groq
Anyscale $0.15 $0.15
Together AI $0.2 $0.2
Replicate $0.05 $0.25
Fireworks $0.2 $0.2
Deepinfra

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

api4all-0.2.4.tar.gz (13.8 kB view details)

Uploaded Source

Built Distribution

api4all-0.2.4-py3-none-any.whl (14.0 kB view details)

Uploaded Python 3

File details

Details for the file api4all-0.2.4.tar.gz.

File metadata

  • Download URL: api4all-0.2.4.tar.gz
  • Upload date:
  • Size: 13.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.9.13

File hashes

Hashes for api4all-0.2.4.tar.gz
Algorithm Hash digest
SHA256 de28229bf06bab9e44ac80ab8ed68f1c4b2f729b2cdb462e264b9fc662fd9b8d
MD5 58b9252ae57cb3185e693cd4f9914ed2
BLAKE2b-256 173fc16a336bfd6d3c6ba4ebd9c7f3256babf4a5adb770d1ab553205751e1aed

See more details on using hashes here.

File details

Details for the file api4all-0.2.4-py3-none-any.whl.

File metadata

  • Download URL: api4all-0.2.4-py3-none-any.whl
  • Upload date:
  • Size: 14.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.9.13

File hashes

Hashes for api4all-0.2.4-py3-none-any.whl
Algorithm Hash digest
SHA256 69b2c5563731e4ecbcab85dd6c5772bce3a9d2ac6d6c5becba36d713d76aed95
MD5 13f928a477a2adfe342e9e217cc15d67
BLAKE2b-256 7483e721ebd74498c862652a3263aaa8f78c8a0c1baab86fea5c053681cf8361

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page