Temporal-Shift-Module icon indicating copy to clipboard operation
Temporal-Shift-Module copied to clipboard

Have you reproduced the accuracy on Something dataset?

Open nemonameless opened this issue 6 years ago • 10 comments

nemonameless avatar Dec 29 '18 05:12 nemonameless

It is still on the way. I will publish my results as soon as I make it.

PingchuanMa avatar Jan 02 '19 10:01 PingchuanMa

@Pika7ma Hi, have you first pre-train on Kinetics ? Otherwise, it seems that the accuracy cannot be reproduced. I just used imagenet-pretrained resnet50 and fixed the bug in https://github.com/Pika7ma/Temporal-Shift-Module/blob/master/tsm_util.py#L14 but only got 41 on something-v1. How about you?

nemonameless avatar Jan 10 '19 07:01 nemonameless

@nemonameless Yes it should be pre-trained on Kinetics following the steps of original paper.

PingchuanMa avatar Jan 10 '19 10:01 PingchuanMa

@Pika7ma did you reproduce kinetics top1 72.5% at 16Frames? I followed tsn based source code(a little diffent at init weight). trained 45 epoch with tran param from paper. But only get 70.436% top1.

Haijunlv avatar Jan 15 '19 01:01 Haijunlv

@Haijunlv Actually no. I am currently doing experiments using 8-frames. I'll try it soon thanku for the information.

PingchuanMa avatar Jan 15 '19 07:01 PingchuanMa

@Pika7ma did you reproduce kinetics top1 72.5% at 16Frames? I followed tsn based source code(a little diffent at init weight). trained 45 epoch with tran param from paper. But only get 70.436% top1.

Hi, I'm interested in that do you use pre-trained resnet50 on ImageNet as init params? In the paper they don't claim for this but I guess it is important.

tzzcl avatar Jan 31 '19 15:01 tzzcl

@tzzcl yes. I use tsn codebase to reproduce. But still have 2% gap with the paper. Maybe pretrained weight is important. But lateset paper by kaiming said pretrained is not necessary. Anyway pretained model save time to converge.

Haijunlv avatar Feb 04 '19 01:02 Haijunlv

@tzzcl yes. I use tsn codebase to reproduce. But still have 2% gap with the paper. Maybe pretrained weight is important. But lateset paper by kaiming said pretrained is not necessary. Anyway pretained model save time to converge.

Hi, I also use the tsn codebase and pretrained resnet50 to reimplement K=8 results on Kinetics. But there is also 2% gap (about 68.8% with my own implementation). Maybe we can only wait for the official code.

tzzcl avatar Feb 04 '19 01:02 tzzcl

@nemonameless May I ask what are your configurations to get 41% on something v1 ? e.g. Batch size, learning rate schedule

gongbudaizhe avatar Feb 25 '19 13:02 gongbudaizhe

hi guys, have you reproduced the accuracy on the Something-Something dataset based on the official code at https://github.com/mit-han-lab/temporal-shift-module ?

sKamiJ avatar Sep 01 '19 07:09 sKamiJ