OFA
OFA copied to clipboard
finetuning with caption_cn_large
I wonder the mini batch size when finetuning with caption_cn_large model on a singe GPU like V100/A100(32G/40G) for caption task. I found that the max batch size is 2/4 on V100/A100. Is there any ways to increase the batch size? fp16 is already set
Try gradient accumulation with --update-freq
Try gradient accumulation with
--update-freq
thanks for suggestion. max mini batch size is 2/4 on V100/A100 for caption task, does this make sense?