OFA icon indicating copy to clipboard operation
OFA copied to clipboard

Config Mismatch between the provided script and pretrained model

Open zhaoweicai opened this issue 2 years ago • 2 comments

Hello, I had a look at the configs in the pretrained OFA-large model, and found a couple of differences. For example:

  • lr = 0.00005 instead of 0.0001 in the script nor 0.0002 in the paper.
  • max_update = 150K, instead of 500K in the paper nor 50 epochs in the script.
  • patch_image_size = 480 instead of 384 in the paper/script.
  • sample_patch_num = -1 instead of 196 in the script.

How was the large model being trained? Is the global_batch_size=batch_size * #GPUs * update_freq = 2048 for all models?

Thanks!

zhaoweicai avatar Jan 09 '23 04:01 zhaoweicai

@zhaoweicai Please refer to Appendix A.2 in our paper. For OFA-Large, we first train it with images of the resolution of 384 × 384 (sample 196 patches), and continue pretraining with images of the resolution of 480 × 480.

logicwong avatar Jan 12 '23 10:01 logicwong

Oh, I was reading the old version of the paper. But still some details are missing for the two-stage training of OFA-Large. I can find the configs of the 2nd stage from the model, but what are parameters for the 1st stage, e.g. LR and max_update? And the datasets are exactly the same for those two stages?

zhaoweicai avatar Jan 26 '23 22:01 zhaoweicai