Skip to main content

dstack is an open-source orchestration engine for running AI workloads on any cloud or on-premises.

Project description

dstack is an open-source container orchestration engine for AI. It accelerates the development, training, and deployment of AI models, and simplifies the management of clusters.

Cloud and on-prem

dstack is easy to use with any cloud or on-prem servers. Supported cloud providers include AWS, GCP, Azure, OCI, Lambda, TensorDock, Vast.ai, RunPod, and CUDO. For using dstack with on-prem servers, see fleets.

Accelerators

dstack supports NVIDIA GPU and Google Cloud TPU out of the box.

Major news ✨

Installation

Before using dstack through CLI or API, set up a dstack server.

Install the server

The easiest way to install the server, is via pip:

pip install "dstack[all]" -U

Configure backends

If you have default AWS, GCP, Azure, or OCI credentials on your machine, the dstack server will pick them up automatically.

Otherwise, you need to manually specify the cloud credentials in ~/.dstack/server/config.yml.

See the server/config.yml reference for details on how to configure backends for all supported cloud providers.

Start the server

To start the server, use the dstack server command:

$ dstack server

Applying ~/.dstack/server/config.yml...

The admin token is "bbae0f28-d3dd-4820-bf61-8f4bb40815da"
The server is running at http://127.0.0.1:3000/

Note It's also possible to run the server via Docker.

Add on-prem servers

If you'd like to use dstack to run workloads on your on-prem servers, see on-prem fleets command.

How does it work?

1. Define run configurations

dstack supports three types of run configurations:

  • Dev environments — for interactive development using a desktop IDE
  • Tasks — for any kind of batch jobs or web applications (supports distributed jobs)
  • Services— for production-grade deployment (supports auto-scaling and authorization)

Each type of run configuration allows you to specify commands for execution, required compute resources, retry policies, auto-scaling rules, authorization settings, and more.

Configuration can be defined as YAML files within your repo.

2. Run configurations

Run any defined configuration either via dstack CLI or API.

dstack automatically handles provisioning, interruptions, port-forwarding, auto-scaling, network, volumes, run failures, out-of-capacity errors, and more.

3. Manage fleets

Use fleets to provision and manage clusters and instances, both in the cloud and on-prem.

More information

For additional information and examples, see the following links:

Contributing

You're very welcome to contribute to dstack. Learn more about how to contribute to the project at CONTRIBUTING.md.

License

Mozilla Public License 2.0

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

dstack-0.18.7.tar.gz (308.5 kB view details)

Uploaded Source

Built Distribution

dstack-0.18.7-py3-none-any.whl (467.1 kB view details)

Uploaded Python 3

File details

Details for the file dstack-0.18.7.tar.gz.

File metadata

  • Download URL: dstack-0.18.7.tar.gz
  • Upload date:
  • Size: 308.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.9.19

File hashes

Hashes for dstack-0.18.7.tar.gz
Algorithm Hash digest
SHA256 0c3d13d3c5a83f959ce986cee1b111cb6b4ab56ba84e6678d79824da59fe1730
MD5 da7a9bbef53948f45c2369ee4ca990ed
BLAKE2b-256 4934e9e1d56b58767c8f9b4984bfc097d657502addd15041a5df495f604ac573

See more details on using hashes here.

File details

Details for the file dstack-0.18.7-py3-none-any.whl.

File metadata

  • Download URL: dstack-0.18.7-py3-none-any.whl
  • Upload date:
  • Size: 467.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.9.19

File hashes

Hashes for dstack-0.18.7-py3-none-any.whl
Algorithm Hash digest
SHA256 6883b42a59c70c125ccbddae7c256645195a4fba05489a50459adc4ee0e8d576
MD5 d11e991491495de942dfca943f8daaf2
BLAKE2b-256 b3d9dea9d02f224c5037c288981b16af6ad865f238799b8f0749732fd04a0588

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page