Skip to main content

No project description provided

Project description

PyPI - Version PyPI - Python Version Static Badge

Unit Testing E2E Testing Coverage Status

🦙 Llama Deploy 🤖

Llama Deploy (formerly llama-agents) is an async-first framework for deploying, scaling, and productionizing agentic multi-service systems based on workflows from llama_index. With Llama Deploy, you can build any number of workflows in llama_index and then run them as services, accessible through a HTTP API by a user interface or other services part of your system.

The goal of Llama Deploy is to easily transition something that you built in a notebook to something running on the cloud with the minimum amount of changes to the original code, possibly zero. In order to make this transition a pleasant one, you can interact with Llama Deploy in two ways:

Both the SDK and the CLI are part of the Llama Deploy Python package. To install, just run:

pip install llama_deploy

[!TIP] For a comprehensive guide to Llama Deploy's architecture and detailed descriptions of its components, visit our official documentation.

Why Llama Deploy?

  1. Seamless Deployment: It bridges the gap between development and production, allowing you to deploy llama_index workflows with minimal changes to your code.
  2. Scalability: The microservices architecture enables easy scaling of individual components as your system grows.
  3. Flexibility: By using a hub-and-spoke architecture, you can easily swap out components (like message queues) or add new services without disrupting the entire system.
  4. Fault Tolerance: With built-in retry mechanisms and failure handling, Llama Deploy adds robustness in production environments.
  5. State Management: The control plane manages state across services, simplifying complex multi-step processes.
  6. Async-First: Designed for high-concurrency scenarios, making it suitable for real-time and high-throughput applications.

[!NOTE] This project was initially released under the name llama-agents, but the introduction of Workflows in llama_index turned out to be the most intuitive way for our users to develop agentic applications. We then decided to add new agentic features in llama_index directly, and focus Llama Deploy on closing the gap between local development and remote execution of agents as services.

Getting Started

The fastest way to start using Llama Deploy is playing with a practical example. This repository contains a few applications you can use as a reference:

We recommend to start from the Quick start example and move to Use a deployment from a web-based user interface immediately after. Each folder contains a README file that will guide you through the process.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

llama_deploy-0.3.1.tar.gz (63.9 kB view details)

Uploaded Source

Built Distribution

llama_deploy-0.3.1-py3-none-any.whl (100.2 kB view details)

Uploaded Python 3

File details

Details for the file llama_deploy-0.3.1.tar.gz.

File metadata

  • Download URL: llama_deploy-0.3.1.tar.gz
  • Upload date:
  • Size: 63.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.8.4 CPython/3.11.0 Linux/6.5.0-1025-azure

File hashes

Hashes for llama_deploy-0.3.1.tar.gz
Algorithm Hash digest
SHA256 a87de2a2cef5016dfe9a2e1ad013f3bd2fbf15c529568177c30c542445e6d7d7
MD5 7ca39b85346b4b913e24f95de403691e
BLAKE2b-256 e7466274813b4325ac94164a9791ce7eb7dd54899af2cf78fe1e3535cfdc86fa

See more details on using hashes here.

File details

Details for the file llama_deploy-0.3.1-py3-none-any.whl.

File metadata

  • Download URL: llama_deploy-0.3.1-py3-none-any.whl
  • Upload date:
  • Size: 100.2 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.8.4 CPython/3.11.0 Linux/6.5.0-1025-azure

File hashes

Hashes for llama_deploy-0.3.1-py3-none-any.whl
Algorithm Hash digest
SHA256 71f3c6ac6715abf3712dc8d01b62932155ff36708c0d0477eefb303897273905
MD5 c1bd568c1096e394dfde33e368ca3b66
BLAKE2b-256 eac89f77f7fdce2b3ed8645efcf2a1d2e6bb8c94e6c9cbb4ef430caf8145c34f

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page