Skip to main content

Easy-to-use LLM API from a state-of-the-art provider and comparison

Project description

api4all

Easy-to-use LLM API from a state-of-the-art provider and comparison.

Installation

  1. Clone the repository
git clone https://github.com/api4all/api4all.git
  1. Create and activate a virtual environment
  • Unix / macOS
python3 -m venv venv
source venv/bin/activate
  • Windows
python3 -m venv venv
.\venv\Scripts\activate
  1. Install the requirements
pip install -r requirements.txt

Providers and Pricing

Providers

Provider Free Credit Rate Limit API Key name Provider string name
Groq Unlimited 30 Requests / Minute GROQ_API_KEY "groq"
Anyscale $10 30 Requests / Second ANYSCALE_API_KEY "anyscale"
Together AI $25 1 Requests / Second TOGETHER_API_KEY "together"
Replicate Free to try 50 Requests / Second REPLICATE_API_KEY "replicate"
Fireworks $1 600 Requests / Minute FIREWORKS_API_KEY "fireworks"
Deepinfra Free to try 200 Concurrent request DEEPINFRA_API_KEY "deepinfra"
Google AI (Vertex AI) Unlimited 60 Requests / Minute GOOGLE_API_KEY "google"
OpenAI 60 Requests / Minute OPENAI_API_KEY "openai"
Mistral AI 5 Requests / Second MISTRAL_API_KEY "mistral"
Anthropic 5 Requests / Minute ANTHROPIC_API_KEY "anthropic"
  • Free to try: Free to try, no credit card required but limited to a certain number of tokens.

Open-source models

1. Mixtral-8x7b-Instruct-v0.1

Mixtral 8x7B is a high-performing sparse mixture of experts model (SMoE) with open weights, licensed under Apache 2.0. It surpasses Llama 2 70B in most benchmarks, offering 6x faster inference. It's the best open-weight model considering cost/performance trade-offs, matching or exceeding GPT3.5 on most standard benchmarks.

  • Developer: Mistral AI
  • Context Length: 32,768
Provider Input Pricing ($/1M Tokens) Output Pricing ($/1M Tokens)
Groq $0 $0
Anyscale $0.5 $0.5
Together AI $0.6 $0.6
Replicate $0.3 $1
Fireworks $0.5 $0.5
Deepinfra $0.27 $0.27
2. Gemma 7B it
  • Developer: Google AI
  • Context Length: 8,192
Provider Input Pricing ($/1M Tokens) Output Pricing ($/1M Tokens)
Groq $0 $0
Anyscale $0.15 $0.15
Together AI $0.2 $0.2
Replicate
Fireworks $0.2 $0.2
Deepinfra
3. LLaMA2-70b
  • Developer: Meta AI
  • Context Length: 4,096
Provider Input Pricing ($/1M Tokens) Output Pricing ($/1M Tokens)
Groq $0 $0
Anyscale $1.0 $1.0
Together AI $0.9 $0.9
Replicate $0.65 $2.75
Fireworks $0.9 $0.9
Deepinfra $0.7 $0.9
4. Mistral-7B-Instruct-v0.1
  • Developer: Mistral AI
  • Context Length: 4,096
Provider Input Pricing ($/1M Tokens) Output Pricing ($/1M Tokens)
Groq
Anyscale $0.15 $0.15
Together AI $0.2 $0.2
Replicate $0.05 $0.25
Fireworks $0.2 $0.2
Deepinfra

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

api4all-0.2.6.tar.gz (13.9 kB view details)

Uploaded Source

Built Distribution

api4all-0.2.6-py3-none-any.whl (14.0 kB view details)

Uploaded Python 3

File details

Details for the file api4all-0.2.6.tar.gz.

File metadata

  • Download URL: api4all-0.2.6.tar.gz
  • Upload date:
  • Size: 13.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.9.13

File hashes

Hashes for api4all-0.2.6.tar.gz
Algorithm Hash digest
SHA256 0a0b61cd3b2a8e65425e28d246fde5fdd168532a7b3a4528b8dffe2dd77f879f
MD5 6abdaa3c309beea50eb47b9d7854d034
BLAKE2b-256 b88306b8c0f421c3b6d1e03d2e027eb751d0bcf5ff182d49cab9df3d45c8e255

See more details on using hashes here.

File details

Details for the file api4all-0.2.6-py3-none-any.whl.

File metadata

  • Download URL: api4all-0.2.6-py3-none-any.whl
  • Upload date:
  • Size: 14.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.9.13

File hashes

Hashes for api4all-0.2.6-py3-none-any.whl
Algorithm Hash digest
SHA256 6422be2fe1a53d50850f36ac8ea79873b990c0865656b10f82602f2cf7e92ec6
MD5 de5085a6175313feaa264ec805044eee
BLAKE2b-256 8dfea786cbd0485703c2fe1bfa034ee5520d5b7816c1011f630bf2ce239aebca

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page