g2p-seq2seq icon indicating copy to clipboard operation
g2p-seq2seq copied to clipboard

Proper training stopping without max_steps configuration

Open jonmay opened this issue 7 years ago • 2 comments

What data and parameters were used to build the pretrained model and the model used to generate PRONALSYL 2007 numbers in row "LSTM num_layers=2, size=512" in the table? The table claims 23.3 WER, when I run the pretrained model on the test set of PRONALSYL 2007 I get 22.8, and when I run my own model trained according to the README in the PRONALSYL download (--train cmudict.dic.train --test cmudict.dic.test --num_layers 2 --size 512 --max_steps 0) I get 32.11. Is the data the same for all these? Are there any tricks I am missing? Are random restarts needed?

jonmay avatar Apr 06 '17 23:04 jonmay

You need more steps, something like --max_steps 70000, we are currently working on proper stopping.

nshmyrev avatar Apr 06 '17 23:04 nshmyrev

ah, thanks, i'll give it a try

jonmay avatar Apr 07 '17 00:04 jonmay