Co-DETR
Co-DETR copied to clipboard
How much VRAM is needed to finetune co_dino_5scale_vit_large_coco?
I have 6 4090 GPUs (VRAM = 120GB). However, when I try to finetune the model, it shows "CUDA out of memory" error. How much VRAM is needed to train the ViT backbone model? I want to know how many GPUs you had when you pretrain the model.
We use 56 A100 80G GPUs to pretrain the model. FSDP and deepspeed can help you reduce training memory consumption, please refer to this issue.