text-generation-webui icon indicating copy to clipboard operation
text-generation-webui copied to clipboard

LLaMa wont load when using deepspeed

Open catalpaaa opened this issue 1 year ago • 2 comments

Describe the bug

LLaMa wont load when using deepspeed, it will just get stucked on

Loading llama-7b...

and taking up all the ram before freezing the system.

Is there an existing issue for this?

  • [X] I have searched the existing issues

Reproduction

fellow the tutorial for enable deepspeed, using llama 7b hf.

Screenshot

No response

Logs

only log shown is Loading llama-7b...

System Info

3080ti 12gb on ubuntu 22

catalpaaa avatar Mar 19 '23 09:03 catalpaaa

btw theres 64gb swap but looks like its not used.

catalpaaa avatar Mar 20 '23 14:03 catalpaaa

Also it looks like whenever the model is splite into mutilple bins then stuff just freeze. To me, deepspeed just doesnt wanna use the swap, it takes up all the buff/cache then dose nothing. image

catalpaaa avatar Mar 20 '23 14:03 catalpaaa

Also it looks like whenever the model is splite into mutilple bins then stuff just freeze. To me, deepspeed just doesnt wanna use the swap, it takes up all the buff/cache then dose nothing. image

Have you found a way to force deepspeed to use swap?

cyrcule avatar Apr 15 '23 16:04 cyrcule