dimensionality-driven-learning icon indicating copy to clipboard operation
dimensionality-driven-learning copied to clipboard

different result on training

Open jiangfanhan opened this issue 6 years ago • 9 comments

Dear Xingjun: I have tried your method to train the 12-layer-cnn on CIFAR-10 with 20% noise rate,I also observe the decrease and increase of LID score. but in my experiment. the test accuracy is 88.34% by just using cross-entropy as loss function. (in your paper is 73.12%) for 40% noise rate the test accuracy is 84.88% (65.07% in your paper),the results is even better than the results using your D2L method. the only difference during the training process may lies on the preprocessing of the training image (what I use is described in the loss correction method CVPR'17). I wonder why the difference is so much and whether you can try the preprocessing method as in CVPR'17 and train the network again. Thanks !

jiangfanhan avatar Jun 21 '18 09:06 jiangfanhan

Hi Fanhan, Will have a look when I have time as I am currently busy with travelling. But from your results it seems CE is even better than any other models which is not reasonable to me. It would be helpful if you can provide some details such as the definition of CE, what is the preprocessing, so that I can replicate your results to understand it precisely.

On Thu, 21 Jun 2018 at 17:22, jiangfanhan [email protected] wrote:

Dear Xingjun: I have tried your method to train the 12-layer-cnn on CIFAR-10 with 20% noise rate,I also observe the decrease and increase of LID score. but in my experiment. the test accuracy is 88.34% by just using cross-entropy as loss function. (in your paper is 73.12%) for 40% noise rate the test accuracy is 84.88% (65.07% in your paper),the results is even better than the results using your D2L method. the only difference during the training process may lies on the preprocessing of the training image (what I use is described in the loss correction method CVPR'17). I wonder why the difference is so much and whether you can try the preprocessing method as in CVPR'17 and train the network again. Thanks !

— You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/xingjunm/dimensionality-driven-learning/issues/1, or mute the thread https://github.com/notifications/unsubscribe-auth/AOIQBkfXlds333ojoX-eFD0tx77Zjyejks5t-2XkgaJpZM4Uxs6j .

-- Best regards, Daniel

xingjunm avatar Jun 26 '18 13:06 xingjunm

Hi Daniel, Thank you for your reply. I have test three method in your paper, they are cross-entropy, forward estimation and D2L on CIFAR-10 dataset. with 20% noise, the test accuracy is 88.34% ,89.35% and 88.65% . with 40% noise, the test accuracy is 84.88%, 86.55% and 85.97% respectively. From other research result the DNN is robust to mild noise so the result just using CE loss is slightly degraded in such case, using correction method can not improve quite much, but the improvement can be observed. The result I get seemed to be closer to the result reported in the CVPR'17 paper, and my preprocessing method is the same as this paper. The CE is defined in the standard way for classification task, I do not take any change.

Regards, Jiangfan

jiangfanhan avatar Jun 27 '18 07:06 jiangfanhan

I have the same issue... I can't reproduce the results on the paper using d2l method even though I didn't change the codes... @xingjunm Could you check out the source code?

pokaxpoka avatar Aug 20 '18 19:08 pokaxpoka

Hi pokaxpoka, thank you for your report. I do find reproductivity issue when tested with a fresh run on new devices. I will try to fix it. Can you provide the details of your results? Did it fail to converge with CIFAR-10 and 40%/60% noise rates?

xingjunm avatar Aug 26 '18 08:08 xingjunm

Thanks @xingjunm When I tried your method it fails to converge with CIFAR-10 (40%/60% noise rate) and CIFAR-100. Thanks for your update.

pokaxpoka avatar Aug 27 '18 01:08 pokaxpoka

any update? does anyone replicate the result?

slowbull avatar Sep 19 '18 13:09 slowbull

I have uploaded an old version of the code old_version/d2l_old.zip, can someone test if this version works?

xingjunm avatar Sep 19 '18 14:09 xingjunm

Thank you all for your interest. And sorry that it took me so long to have time fix the issue. Sorry for the waiting. Good luck to you all with your papers.

xingjunm avatar Nov 07 '18 06:11 xingjunm

@jiangfanhan , I did also observe that the accuracy of CCE is much higher than reported in the paper, if early stopping criterion is used with respect to (noisy) validation data. I got CCE accuracy=0.81

bbdamodaran avatar Nov 26 '18 15:11 bbdamodaran