T2T-ViT icon indicating copy to clipboard operation
T2T-ViT copied to clipboard

gained 0.094 for eval_top1, and 0.33 for eval_top5, after 36-epoch training on 8 gpus

Open CheerM opened this issue 3 years ago • 4 comments

hi, would you mind releasing the training log for T2t-vit-t-14 training with 8 GPUs? I tried to rerun the script for training T2t-vit-t-14 with 8 GPUs. It gained 0.094 for eval_top1, 0.33 for eval_top5, after 36 epochs. It seems too slow to converge.

CheerM avatar Jun 01 '21 07:06 CheerM

I don't know about the convergence over time but it should take 310 epochs to get the paper results.

lucastononrodrigues avatar Jun 01 '21 12:06 lucastononrodrigues

hi, would you mind releasing the training log for T2t-vit-t-14 training with 8 GPUs? I tried to rerun the script for training T2t-vit-t-14 with 8 GPUs. It gained 0.094 for eval_top1, 0.33 for eval_top5, after 36 epochs. It seems too slow to converge.

Hi, the log of T2t-vit-t-14 is trained with 8 GPUs. It's normal if your results are slightly higher or lower than the logs.

yuanli2333 avatar Jun 01 '21 15:06 yuanli2333

hi, would you mind releasing the training log for T2t-vit-t-14 training with 8 GPUs? I tried to rerun the script for training T2t-vit-t-14 with 8 GPUs. It gained 0.094 for eval_top1, 0.33 for eval_top5, after 36 epochs. It seems too slow to converge.

Hello, have you solved the problem? I have the same problem. And the loss doesn't decrease.

WangChen0902 avatar Dec 28 '21 06:12 WangChen0902

Same here, training T2t-vit-t-14 on 3 GPU with -b64, and after 80 epochs: top-1 acc = 0.095%, top-5 acc = 0.301%

It seems no improvements from epoch 20 (Top1=0.093, Top5=0.3199) through 80.

imkzh avatar Mar 10 '22 06:03 imkzh