randomNAS_release icon indicating copy to clipboard operation
randomNAS_release copied to clipboard

Code for search & train WITHOUT weight sharing

Open killawhale2 opened this issue 6 years ago • 3 comments

The authors report the results for the random search baseline without any weight sharing but do not provide the code to reproduce the baseline's result. This would be helpful in fully reproducing their tables in the results section and not just their weight-shared versions.

killawhale2 avatar Jun 20 '19 14:06 killawhale2

This code is provided in a repository linked from the README: https://github.com/liamcli/darts_asha

Note that you won't be able to get exactly reproducible results for the search phase due to asynchronous promotions in the parallel setting.

liamcli avatar Jun 20 '19 16:06 liamcli

Thanks for the quick reply. I've checked the repo again, and I still only see code for running random search with ASHA. I'm mentioning this because the paper claims the random search baseline takes 4 gpu days while the random search with ASHA takes 9 gpu days for the cnn experiments. I'd like to reproduce the results for the base random search as that is more computationally viable for my resources.

killawhale2 avatar Jun 21 '19 05:06 killawhale2

The random search baseline that takes 4 GPU days is from the original DARTS paper. You can just use the ASHA codebase with no early-stopping to run random search for as long as you want.

liamcli avatar Jun 21 '19 06:06 liamcli