dstack
dstack copied to clipboard
dstack is an open-source alternative to Kubernetes, designed to simplify development, training, and deployment of AI across any cloud or on-prem. It supports NVIDIA, AMD, and TPU.
### Steps to reproduce 1. Run the "Build docker & cloud images" GitHub Action. Uncheck "Build docker images", check "Staging build" and one of the clouds. 2. Wait until the...
Currently, I have to specify: ``` model: type: chat name: meta-llama/Meta-Llama-3.1-70B-Instruct format: openai ``` I suggest that we allow to specify ``` model: meta-llama/Meta-Llama-3.1-70B-Instruct ``` In that case `type` and...
### Problem In order to use private Docker images hosted with AWS ECR, one has to [generate](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html) a temporarily password using `aws ecr get-login-password` and then hardcode them in the...
### Problem Run logs are never deleted, so disk space usage on the server can grow quickly and indefinitely. ### Solution Add a setting that will specify the TTL for...
### Problem In the server console logs, it's unclear that a run was terminated due to max_duration. Attaching image showing when a run was started and the messaging shown 6...
### Problem Currently, with the AWS backend `dstack` unconditionally requests one network interface, even with instance types that have multiple network cards (e.g., `p5.48xlarge` has 32 EFA-capable cards). Network performance...
### Steps to reproduce 1. make a `repro.dstack.yml` with: ```yaml type: task name: my-repro-task image: dstackai/dind:latest privileged: true commands: - start-dockerd - sleep infinity resources: cpu: 4.. memory: 6GB.. gpu:...
Volumes can be supported via [File Systems](https://lambdalabs.com/blog/persistent-storage-beta). File Systems are persistent disks that can be mounted at fixed mount points of one or more VMs. It’s not possible to use...
### Steps to reproduce Create a fleet with an on-prem node/instance. Change the name of the fleet in the config and apply. The on-prem node will now be a part...