Skip to main content

Open-source, OpenAI-compatible API server with pluggable providers for any model and any infrastructure

Project description

Llama Stack

PyPI Version PyPI Downloads Docker Hub Pulls License Discord Unit Tests Integration Tests OpenResponses Conformance Ask DeepWiki

Quick Start | Documentation | OpenAI API Compatibility | Discord

Open-source agentic API server for building AI applications. OpenAI-compatible. Any model, any infrastructure.

Llama Stack Architecture

Llama Stack is a drop-in replacement for the OpenAI API that you can run anywhere — your laptop, your datacenter, or the cloud. Use any OpenAI-compatible client or agentic framework. Swap between Llama, GPT, Gemini, Mistral, or any model without changing your application code.

from openai import OpenAI

client = OpenAI(base_url="http://localhost:8321/v1", api_key="fake")
response = client.chat.completions.create(
    model="llama-3.3-70b",
    messages=[{"role": "user", "content": "Hello"}],
)

What you get

  • Chat Completions & Embeddings — standard /v1/chat/completions, /v1/completions, and /v1/embeddings endpoints, compatible with any OpenAI client
  • Responses API — server-side agentic orchestration with tool calling, MCP server integration, and built-in file search (RAG) in a single API call (learn more)
  • Vector Stores & Files/v1/vector_stores and /v1/files for managed document storage and search
  • Batches/v1/batches for offline batch processing
  • Open Responses conformant — the Responses API implementation passes the Open Responses conformance test suite

Use any model, use any infrastructure

Llama Stack has a pluggable provider architecture. Develop locally with Ollama, deploy to production with vLLM, or connect to a managed service — the API stays the same.

See the provider documentation for the full list.

Get started

Install and run a Llama Stack server:

# One-line install
curl -LsSf https://github.com/llamastack/llama-stack/raw/main/scripts/install.sh | bash

# Or install via uv
uv pip install llama-stack

# Start the server (uses the starter distribution with Ollama)
llama stack run

Then connect with any OpenAI client — Python, TypeScript, curl, or any framework that speaks the OpenAI API.

See the Quick Start guide for detailed setup.

Resources

Client SDKs:

Language SDK Package
Python llama-stack-client-python PyPI version
TypeScript llama-stack-client-typescript NPM version

Community

We hold regular community calls every Thursday at 09:00 AM PST — see the Community Event on Discord for details.

Star History Chart

Thanks to all our amazing contributors!

Llama Stack contributors

Project details


Release history Release notifications | RSS feed

This version

0.7.1

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

llama_stack-0.7.1.tar.gz (15.9 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

llama_stack-0.7.1-py3-none-any.whl (782.7 kB view details)

Uploaded Python 3

File details

Details for the file llama_stack-0.7.1.tar.gz.

File metadata

  • Download URL: llama_stack-0.7.1.tar.gz
  • Upload date:
  • Size: 15.9 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for llama_stack-0.7.1.tar.gz
Algorithm Hash digest
SHA256 d88dc8430abe1d26f3908ab506f0f6ccd4f3c16ba06ee46ac662f4c896c52da8
MD5 a6d45e1193de4adeda8500c88b8405d9
BLAKE2b-256 849ba8577f0761b02be4d45d625cf5870ab64859dbbe219088d6c8a39e9cb79d

See more details on using hashes here.

Provenance

The following attestation bundles were made for llama_stack-0.7.1.tar.gz:

Publisher: pypi.yml on llamastack/llama-stack

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file llama_stack-0.7.1-py3-none-any.whl.

File metadata

  • Download URL: llama_stack-0.7.1-py3-none-any.whl
  • Upload date:
  • Size: 782.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for llama_stack-0.7.1-py3-none-any.whl
Algorithm Hash digest
SHA256 65b9c827989011af3879ad051db9134163f8e6b3bc3685340328f21f5d07cf5f
MD5 94999741d0921de745061b8e89db9cad
BLAKE2b-256 be41f57aedebba533bf5e8b7c5f3f2ee1edb88409e1aa5c83fa7ddc5f85b3c48

See more details on using hashes here.

Provenance

The following attestation bundles were made for llama_stack-0.7.1-py3-none-any.whl:

Publisher: pypi.yml on llamastack/llama-stack

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page