faster-rnnlm icon indicating copy to clipboard operation
faster-rnnlm copied to clipboard

Benchmark on 1 Billion

Open sherjilozair opened this issue 8 years ago • 5 comments

You report training times in the 1 Billion dataset, but not test perplexity. It would be helpful to compare this implementation with others if you could post the test perplexity and the hyperparameters that could be used to reproduce the result.

Also, thanks for putting this on Github. :)

sherjilozair avatar Aug 13 '15 07:08 sherjilozair

I will definitely add some ASAP! And could you advice any other implementations that could handle one billion word dataset? I'm awawre about torch-based HS. Everything else seems to be too slow for such a huge vocabulary.

akhti avatar Aug 13 '15 14:08 akhti

Possibly the only library that comes close to handling it right now is: https://github.com/nicholas-leonard/dp/.

Looking forward to your results.

sherjilozair avatar Aug 13 '15 21:08 sherjilozair

I've finally added information about finally of different models. Check out https://github.com/yandex/faster-rnnlm#experiments

akhti avatar Sep 10 '15 16:09 akhti

The benchmarks are really cool! Thanks!

How do you interpolate with the KN-5 model, though? And from where is the KN-5 model obtained?

sherjilozair avatar Sep 16 '15 00:09 sherjilozair

The model is downloaded from benchmark's repo. I used quite straightforward interpolation, i.e. mean per-word probability, as it's describe in the benchmarks' paper.

akhti avatar Sep 16 '15 08:09 akhti