Reza Yazdani

Results 95 comments of Reza Yazdani

the results might be different, but is it meaningful when using kernels? Can you please paste the output?

Interesting, I am not sure what might be different between our system environment that we see different results! I am using Torch1.12+CUDA11.6 and I see similar results between HF and...

Can you please paste the whole log? I want to see the transformer configuration

Sorry, I meant the output log when you are running the test.

Hi @AlexWortega, I did retry running this with the same test above. I also did modify it to be similar to yours. However, I am still seeing similar results between...

Hi @AlexWortega , Here is the output of the two commands: ``` ds_report -------------------------------------------------- DeepSpeed C++/CUDA extension op report -------------------------------------------------- NOTE: Ops not installed will be just-in-time (JIT) compiled at...

Hi @reymondzzzz Thanks for the PR. I see this can fix some assumptions we have on model size or batch size during the runtime. But, would you mind give a...

Hi @mallorbc , The problem is that the model selected from HF is Fp32 and it will load the checkpoint before coming to the model-partitioning on the DeepSeed-Inference side. For...

> By changing the pipeline to the following I now get VRAM usage of roughly 12GB per GPU. However, shouldn't the model be split over both GPUs and thus roughly...

Hi @trianxy , I think I know where this issue is coming from. It is due to reducing the max-tokens to 128 [here](https://github.com/microsoft/DeepSpeed/blob/master/csrc/transformer/inference/includes/custom_cuda_layers.h#L20). We have a [PR](https://github.com/microsoft/DeepSpeed/pull/2212) to fix this...