regisss
regisss
Thanks. What is the command you use to run this script?
I cannot reproduce it, it works on my side. Can you provide the full logs of your run and the output of `pip list` please?
Are you using Gaudi1 or Gaudi2?
Can you share the logs of your run please? For a 634M-parameter model, you should be able to fit much bigger batches on Gaudi2.
Hi @rubenCrayon! `deepspeed_reinit` was removed a few versions ago, you should use a more recent version of Optimum. Which may requires to change your script a bit, in that case...
@sgugger All tests passed, so I think this one can be merged :slightly_smiling_face:
@plamb-viso Here is the guide to add ONNX export support for a new architecture in Optimum: https://huggingface.co/docs/optimum/exporters/onnx/usage_guides/contribute Feel free to open a PR there and we'll help you if you...
Hi @hannan72! I recommend that you use Optimum for exporting Whisper to the ONNX format (it will basically be a wrapper around `torch.onnx.export` but it is tested and Whisper is...
Yes I see you opened this issue in Optimum: https://github.com/huggingface/optimum/issues/827 I think the best is to wait for @fxmarty to take a look at it. Regarding these warnings, I don't...
`torchrun` is equivalent to `python -m torch.distributed.run` while `python -m torch.distributed.launch` is deprecated. I think the reason why it is deprecated is just that `torchrun` does the same but also...