DeepSpeed
DeepSpeed copied to clipboard
[REQUEST] M1 Max support
Is your feature request related to a problem? Please describe. I'm looking to do some fine tuning of GPT-J on a Macbook Pro M1 Max w/ 64GB RAM
Describe the solution you'd like Given the CPU, GPU, and memory are significant for this chip it seems a reasonable target and would skip the hassle and cost management/budget asks for cloud based training.
We’ve definitely been watching if/when pytorch will support M1, it sounds like it’s planned though.
https://github.com/pytorch/pytorch/issues/47702
Specifically see this comment: https://github.com/pytorch/pytorch/issues/47702#issuecomment-965625139
In terms of DeepSpeed support for M1 I suspect (depending on final design on torch side) many of our features will work well. However, we’ll have to reassess once torch releases their updated plan and final support here.
Looks like pytorch released support.
any updates?
+1 on the request
+1 for this, would be very helpful to have support from deepspeed in addition to torch.
+1 on the update, it would be very cool to have deepspeed now working on Apple M1/M2 machines now that PyTorch is supporting them..
It is exciting to see PyTorch support for M1/M2. We very open to extending deepspeed support to more and more accelerators but we currently lack bandwidth and hardware to explore this at the moment. However, we would gladly support any PRs for this similar to our ongoing support for the following Intel accelerator PR: https://github.com/microsoft/DeepSpeed/pull/2221
Any update? it's been a long time.
@phnessu4, unfortunately no update here as we have not had bandwidth or hardware access to drive this line. We would gladly accept any PR in this direction. Our accelerator abstraction is completed, and Intel XPU is now fully supported.