video_reloc icon indicating copy to clipboard operation
video_reloc copied to clipboard

questions about the performance

Open newdpdeeplearner opened this issue 4 years ago • 4 comments

excuse me, i've read your paper which is impressive. so i run your code on my own machine. While the result of the proposed model is reproducible, the results of the 'Chance' baseline are much worse as comparted to those reported in your paper. Specifically, the random temporal boundaries were generated through 'start = tf.random_uniform([], maxval=length, dtype=tf.int32), end = tf.random_uniform([], minval=start, maxval=length+1, dtype=tf.int32)' where 'length' is the mapped video_length of 'len_r' in your code. However, the performance are as follows:

Saving dict for global step 200: IoU/0.1 = 0.42229038, IoU/0.2 = 0.32719836, IoU/0.3 = 0.23721881, IoU/0.4 = 0.17177914, IoU/0.5 = 0.111451946, IoU/0.6 = 0.06543967, IoU/0.7 = 0.03578732, IoU/0.8 = 0.020449897, IoU/0.9 = 0.0040899795, IoU/mean = 0.047443762

i wonder whether my implementation is right or not. Thank you

newdpdeeplearner avatar Jun 04 '20 03:06 newdpdeeplearner

Could you first sample length uniformly and then sample the start position?

fengyang0317 avatar Jun 04 '20 03:06 fengyang0317

i'm sorry, what does 'sample length uniformly' mean ? i mean, isn't the length of each reference video is a fixed number? or you mean i should calculate the average length of all the moments in the dataset, and then sample the start position and determine the video segment with the average length?

newdpdeeplearner avatar Jun 04 '20 04:06 newdpdeeplearner

If the length of the reference is L, sample the target length in the range [1, L] uniformly.

fengyang0317 avatar Jun 04 '20 16:06 fengyang0317

thank you very much!

newdpdeeplearner avatar Jun 05 '20 01:06 newdpdeeplearner