Skip to main content

Llama Stack

Project description

Llama Stack Logo

Llama Stack

PyPI version PyPI - Downloads Discord

Get Started | Documentation

This repository contains the Llama Stack API specifications as well as API Providers and Llama Stack Distributions.

The Llama Stack defines and standardizes the building blocks needed to bring generative AI applications to market. These blocks span the entire development lifecycle: from model training and fine-tuning, through product evaluation, to building and running AI agents in production. Beyond definition, we are building providers for the Llama Stack APIs. These were developing open-source versions and partnering with providers, ensuring developers can assemble AI solutions using consistent, interlocking pieces across platforms. The ultimate goal is to accelerate innovation in the AI space.

The Stack APIs are rapidly improving, but still very much work in progress and we invite feedback as well as direct contributions.

APIs

The Llama Stack consists of the following set of APIs:

  • Inference
  • Safety
  • Memory
  • Agentic System
  • Evaluation
  • Post Training
  • Synthetic Data Generation
  • Reward Scoring

Each of the APIs themselves is a collection of REST endpoints.

API Providers

A Provider is what makes the API real -- they provide the actual implementation backing the API.

As an example, for Inference, we could have the implementation be backed by open source libraries like [ torch | vLLM | TensorRT ] as possible options.

A provider can also be just a pointer to a remote REST service -- for example, cloud providers or dedicated inference providers could serve these APIs.

Llama Stack Distribution

A Distribution is where APIs and Providers are assembled together to provide a consistent whole to the end application developer. You can mix-and-match providers -- some could be backed by local code and some could be remote. As a hobbyist, you can serve a small model locally, but can choose a cloud provider for a large model. Regardless, the higher level APIs your app needs to work with don't need to change at all. You can even imagine moving across the server / mobile-device boundary as well always using the same uniform set of APIs for developing Generative AI applications.

Supported Llama Stack Implementations

API Providers

API Provider Builder Environments Agents Inference Memory Safety Telemetry
Meta Reference Single Node :heavy_check_mark: :heavy_check_mark: :heavy_check_mark: :heavy_check_mark: :heavy_check_mark:
Fireworks Hosted :heavy_check_mark: :heavy_check_mark: :heavy_check_mark:
AWS Bedrock Hosted :heavy_check_mark: :heavy_check_mark:
Together Hosted :heavy_check_mark: :heavy_check_mark: :heavy_check_mark:
Ollama Single Node :heavy_check_mark:
TGI Hosted and Single Node :heavy_check_mark:
Chroma Single Node :heavy_check_mark:
PG Vector Single Node :heavy_check_mark:
PyTorch ExecuTorch On-device iOS :heavy_check_mark: :heavy_check_mark:

Distributions

Distribution Llama Stack Docker Start This Distribution Inference Agents Memory Safety Telemetry
Meta Reference llamastack/distribution-meta-reference-gpu Guide meta-reference meta-reference meta-reference; remote::pgvector; remote::chromadb meta-reference meta-reference
Meta Reference Quantized llamastack/distribution-meta-reference-quantized-gpu Guide meta-reference-quantized meta-reference meta-reference; remote::pgvector; remote::chromadb meta-reference meta-reference
Ollama llamastack/distribution-ollama Guide remote::ollama meta-reference remote::pgvector; remote::chromadb meta-reference meta-reference
TGI llamastack/distribution-tgi Guide remote::tgi meta-reference meta-reference; remote::pgvector; remote::chromadb meta-reference meta-reference
Together llamastack/distribution-together Guide remote::together meta-reference remote::weaviate meta-reference meta-reference
Fireworks llamastack/distribution-fireworks Guide remote::fireworks meta-reference remote::weaviate meta-reference meta-reference

Installation

You have two ways to install this repository:

  1. Install as a package: You can install the repository directly from PyPI by running the following command:

    pip install llama-stack
    
  2. Install from source: If you prefer to install from the source code, follow these steps:

     mkdir -p ~/local
     cd ~/local
     git clone git@github.com:meta-llama/llama-stack.git
    
     conda create -n stack python=3.10
     conda activate stack
    
     cd llama-stack
     $CONDA_PREFIX/bin/pip install -e .
    

Documentations

Please checkout our Documentations page for more details.

Llama Stack Client SDK

Language Client SDK Package
Python llama-stack-client-python PyPI version
Swift llama-stack-client-swift Swift Package Index
Node llama-stack-client-node NPM version
Kotlin llama-stack-client-kotlin Maven version

Check out our client SDKs for connecting to Llama Stack server in your preferred language, you can choose from python, node, swift, and kotlin programming languages to quickly build your applications.

You can find more example scripts with client SDKs to talk with the Llama Stack server in our llama-stack-apps repo.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

llama_stack-0.0.53.tar.gz (197.0 kB view details)

Uploaded Source

Built Distribution

llama_stack-0.0.53-py3-none-any.whl (356.2 kB view details)

Uploaded Python 3

File details

Details for the file llama_stack-0.0.53.tar.gz.

File metadata

  • Download URL: llama_stack-0.0.53.tar.gz
  • Upload date:
  • Size: 197.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.10.15

File hashes

Hashes for llama_stack-0.0.53.tar.gz
Algorithm Hash digest
SHA256 0c0fc430ced6ed973b405fa65a36acea3567dee3875f17b4612e8067cd0173a0
MD5 608f8389ac3bb5e44e9999b64df15888
BLAKE2b-256 997c886a1b5ac0025d72d27e18d11582804528cc192c7e7067d83208a2a9884e

See more details on using hashes here.

File details

Details for the file llama_stack-0.0.53-py3-none-any.whl.

File metadata

  • Download URL: llama_stack-0.0.53-py3-none-any.whl
  • Upload date:
  • Size: 356.2 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.10.15

File hashes

Hashes for llama_stack-0.0.53-py3-none-any.whl
Algorithm Hash digest
SHA256 f98409b45d6b1c3c4968289b429d650bfaba5333b3f0b0148e09b0dcada7461d
MD5 dea59c8ee7d25f640b04c93dd58cb259
BLAKE2b-256 ececbc9a33acb0f6bfd36044c9f15d0f8acfa81b1401b1df953979f67a2fc66a

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page