Text-To-Video-Finetuning icon indicating copy to clipboard operation
Text-To-Video-Finetuning copied to clipboard

First GPU occupies more VRAM in distributed training

Open suzhenghang opened this issue 2 years ago • 0 comments
trafficstars

link, device = torch.device("cuda" if torch.cuda.is_available() else "cpu") cached_latent = torch.load(self.cached_data_list[index], map_location=device) Otherwise, in multi-GPU distributed training, the first GPU may occupy excessive VRAM compared to the other GPUs.

suzhenghang avatar May 22 '23 07:05 suzhenghang