DSS-pytorch icon indicating copy to clipboard operation
DSS-pytorch copied to clipboard

result compare

Open KunWangV opened this issue 6 years ago • 11 comments

Have you compared the results with paper?

KunWangV avatar May 09 '18 08:05 KunWangV

Sorry, I didn't compare with paper. (As I know, the author release the dataset now, you can train it by yourself ) --- I am not working on it recently, so i am sorry ~

AceCoooool avatar May 09 '18 11:05 AceCoooool

Any news on that? :)

eduardramon avatar Sep 09 '18 16:09 eduardramon

so,does anyone compare the result of this work with the paper?

holyhao avatar Sep 21 '18 06:09 holyhao

I will do it in October. (I am finding a job recently, so forgive me.)

AceCoooool avatar Sep 21 '18 06:09 AceCoooool

emm,i just want to make sure if this repo works well。

holyhao avatar Sep 22 '18 08:09 holyhao

@holyhao I have update the code and results in pre-version,I will update results of v2(using learnable fusion) tomorrow

AceCoooool avatar Oct 12 '18 10:10 AceCoooool

@holyhao I have update the code and results in pre-version,I will update results of v2(using learnable fusion) tomorrow

Thanks for your work. By the way,do you train the net on larger batchsize and lr. I see that you set batchszie only 8 and lr is even small.

holyhao avatar Oct 17 '18 02:10 holyhao

  1. I think it's better to use larger lr (I find the loss curve decrease too slow at the beginning). --- but I did not try it (the learning in the paper is very small --- 1e-8)
  2. You can try larger batchsize(and can also use vgg with batch normalization --- I use pre-trained vgg without bn layer, and use a pre-trained model from caffe may increase the results, many projects find caffe-pretrained model have good performance). welcome to share your results, thank you ~

AceCoooool avatar Oct 17 '18 02:10 AceCoooool

I try lr=1e-4 with backbone resnet18 ,it works fine and converges faster. But when i try larger batchsize like 32,48, it converges slow and gets worse val results. It confused me, as far as I known, larger batchsize should leads to better results. Do you have some ideas about this.

holyhao avatar Oct 17 '18 02:10 holyhao

I did not have a machine with large memory(GPU), so i have no "engineering experiment" in large batch size. However, there is several discuss about bach-size:

  1. stack exchange
  2. 《Deep Learning》Ian p172 (Chinese version):small batch size may have regularization, with better generalization.

However, I think there must be some practical tip to train model in larger batch size. (I am sorry without good suggestions. :cry: )

AceCoooool avatar Oct 17 '18 03:10 AceCoooool

Your reply really inspires me and thank you very much.

holyhao avatar Oct 17 '18 06:10 holyhao