e5-mistral-7b-instruct
e5-mistral-7b-instruct copied to clipboard
OOM with 2 GPUs (48GB in total)
Hi, i do not understand why the execution fails when I use 2 GPUs, however with a single one it works.
I tried the following setups:
- Deafult configuration (i just changed
gradient_accumulation_stepsto 1), it works - Default configuration (
gradient_accumulation_steps=1andnum_processes=2), torch.cuda.OutOfMemoryError
I am using docker, hardware: 2xL4 (24GBx2)./