dstack is an open-source orchestration engine for running AI workloads in any cloud or data center.
10K+
dstack is a streamlined alternative to Kubernetes, specifically designed for AI. It simplifies container orchestration
for AI workloads both in the cloud and on-prem, speeding up the development, training, and deployment of AI models.
dstack supports NVIDIA GPU, AMD GPU, and Google Cloud TPU out of the box.
To use dstack with your own cloud accounts, create the ~/.dstack/server/config.yml file and
configure backends.
Starting the dstack server via Docker can be done the following way:
docker run -p 3000:3000 -v $HOME/.dstack/server/:/root/.dstack/server dstackai/dstack
The dstack server is running at http://0.0.0.0:3000
The admin user token is 'bbae0f28-d3dd-4820-bf61-8f4bb40815da'
For more details on server configuration options, see the server deployment guide.
To point the CLI to the dstack server, configure it
with the server address, user token, and project name:
$ pip install dstack
$ dstack project add --name main \
--url http://127.0.0.1:3000 \
--token bbae0f28-d3dd-4820-bf61-8f4bb40815da
Configuration is updated at ~/.dstack/config.yml
If you want the dstack server to run containers on your on-prem servers,
use fleets.
For additional information and examples, see the following links:
Content type
Image
Digest
sha256:2047fbcdb…
Size
265.4 MB
Last updated
4 days ago
Requires Docker Desktop 4.37.1 or later.