DeepSpeedExamples
DeepSpeedExamples copied to clipboard
Can I train a opt-6.7B model on 4x4090 gpus?
I get 4 4090 gpus, and I want to train the opt-6.7B by using DeepSpeed Chat. Is that possible? I mean I have no idea if I should get a GPU that has enough VRAM or maybe I can use 4 small gpus instead.
I have the same question
How large memory does 4090 have? We have basic instruction in the tutorial/readme of step 3
How large memory does 4090 have? We have basic instruction in the tutorial/readme of step 3
One single 4090 has 24Gb VRAM, so 4 4090s have a total of 96Gb. Is it possible to train the opt-6.7B?
Looks like the memory is enough. Can you share the training script you are using for now?
Close as no followup