Michael Luo
Michael Luo
+1 with Romil's idea. This aligns with the ultimate goal to make the remote env as similar as possible as to the local env.
Yes, will fix, thanks for the catch @WoosukKwon !
@WoosukKwon `delete_remote_dir` is similar to deleting a bucket, in `_delete_s3_bucket` or `_delete_gcs_bucket` in `S3Store` and `GCSStore` respectively. Do you want an option to delete a subdir within the bucket?
It looks like for Justin, the version of Ray doesn't matter.
Yes, this was launching two jobs with `custom_resource={"V100": 2}`.
Verified on master.
Admins only use `ray status` after they set up Ray cluster and to check if it is running. Otherwise, I think this fix works.
@romilbhardwaj @concretevitamin Is this still an issue atm?
The fix is `(ps aux | grep -v bash | grep -v grep | grep -q sky.skylet.skylet) || (nohup python3 -m sky.skylet.skylet >> ~/.sky/skylet.log 2>&1 &) `