VaDE-pytorch icon indicating copy to clipboard operation
VaDE-pytorch copied to clipboard

hyper-parameters

Open MiguelMonteiro opened this issue 5 years ago • 10 comments

Hello,

Thanks for sharing this code. I was wondering if you could also share the hyper-parametrization you have used in order to obtain the training curve on the front page: 94% cluster accuracy after 300 epochs. Running the code out of the box it seems to plateau at around 80%

Thank you,

Best regards,

Miguel

MiguelMonteiro avatar Jan 23 '20 18:01 MiguelMonteiro

I have the same problem!!!

Jillian555 avatar Mar 05 '20 07:03 Jillian555

I just run code it can get the 94%+ accuracy..

YUMI66666 avatar Apr 05 '20 05:04 YUMI66666

I meet the same problem, and just setting --hid_dim 50 can obtain 94%+ acc

zhangzs-pku avatar May 31 '20 07:05 zhangzs-pku

I meet the same problem, and just setting --hid_dim 50 can obtain 94%+ acc

I set the hid_dim as 50, get worse acc, any ideas?

goodman1204 avatar Jul 04 '20 02:07 goodman1204

The variance of the method is quite high. There is no way to avoid this. Try to repeat the experiment at least 10 times.

wzm2256 avatar Jul 13 '20 02:07 wzm2256

This method relies on the GMM to initialize the network, and you can get the final result by trying a few more times!

GuHongyang avatar Jul 13 '20 09:07 GuHongyang

@GuHongyang Yes, but that is unfair. Because in practice you simply do not have any label to evaluate your results.

wzm2256 avatar Jul 13 '20 09:07 wzm2256

Yes. The core problem is still the algorithm itself. You can divide the training data into a part of the data as validation data, and use this data to find better GMM initialization parameters.

GuHongyang avatar Jul 13 '20 09:07 GuHongyang

The variance of the method is quite high. There is no way to avoid this. Try to repeat the experiment at least 10 times.

Thank for your comments. I am wondering if the accuracy will be infected if I don't use GPU ( I removed the cuda() part).

goodman1204 avatar Jul 16 '20 03:07 goodman1204

Yes. The core problem is still the algorithm itself. You can divide the training data into a part of the data as validation data, and use this data to find better GMM initialization parameters.

Thanks for the advice. I have the same problem, I wonder if the network is correctly initialized. I ran the code for about 10 times with PyTorch 1.7, but it can never achieve 82%+. Is there any difference between different PyTorch versions in network initialization? Besides, can you share the final code in this repository?

RenShijie avatar Dec 03 '20 20:12 RenShijie