uda icon indicating copy to clipboard operation
uda copied to clipboard

Hyper parameters about 'uda_softmax_temp' and 'uda_confidence_thresh'

Open Lan1991Xu opened this issue 5 years ago • 5 comments

Hi, Many thanks for sharing the code. Could please tell me when the labelled data is limited, e,p CIFAR-10 250 labels, the hyperparameters of 'uda_softmax_temp' and 'uda_confidence_thresh'. Thanks

Lan1991Xu avatar Jul 19 '19 18:07 Lan1991Xu

Hi, we set uda_confidence_thresh to 0.8 and uda_softmax_temp to 0.9 for that case. Thanks!

michaelpulsewidth avatar Jul 21 '19 05:07 michaelpulsewidth

Hi @qizhex,

Are those the hyper-parameters that give SOTA performance (i.e. error rate of 8.41%) on 250 labels for CIFAR-10?

Also, what were the hyperparameters of 'uda_softmax_temp' and 'uda_confidence_thresh' used for 500, 1000, 2000, and 4000 labels for CIFAR-10?

I noticed that in #14, the command to run the SOTA-run for CIFAR-10 on 4000 labels didn't use 'uda_confidence_thresh' to achieve the best results, so I'm curious at what point 'uda_confidence_thresh' no longer provided benefits to the training process.

Thanks!

varunnair18 avatar Jul 31 '19 08:07 varunnair18

Hi, we used uda_softmax_temp and uda_confidence_thresh only for 250 labels. When the size of the labeled set is 500, 1000, 2000, we simply use entropy minimization with coefficient 0.1. For CIFAR-10, the entropy should be computed on original unlabeled examples instead of augmented unlabeled examples. For ImageNet, using original examples or augmented examples for entropy minimization leads to similar performance, so in a previous version of the code, the entropy was computed on augmented unlabeled examples.

michaelpulsewidth avatar Aug 02 '19 02:08 michaelpulsewidth

Got it, thank you for the information!

varunnair18 avatar Aug 02 '19 12:08 varunnair18

Thank you!

guotong1988 avatar Jun 05 '20 04:06 guotong1988