practical_seq2seq icon indicating copy to clipboard operation
practical_seq2seq copied to clipboard

Too many UNK in the output

Open LeenaShekhar opened this issue 7 years ago • 2 comments

I trained the network for 5000 iterations; see the loss below

Building Graph Training started

Model saved to disk at iteration #1000 val loss : 3.290424

Model saved to disk at iteration #2000 val loss : 3.261373

Model saved to disk at iteration #3000 val loss : 3.224990

Model saved to disk at iteration #4000 val loss : 3.151570

Model saved to disk at iteration #5000 val loss : 3.155647

After this I wanted to evaluate the model on the test dataset. Most of the decoder's output is "unk" (see below):

q : [hillary is crazy also evil nothing good about her except that she has a terminal illness]; a : [i unk unk unk unk unk unk unk unk unk unk unk unk unk] q : [breaking unk unk israeli unk and unk peace prize winner dies at unk]; a : [unk unk unk unk unk unk unk unk unk unk unk unk unk] q : [because and jason unk are fighting in the cage next week to see who unk into whom]; a : [i unk unk unk unk unk unk unk unk unk unk unk unk unk] q : [im considering unk a ticket shit looks live ]; a : [i unk unk unk] q : [unk is a classic but tears in heaven is stupid]; a : [i unk unk unk unk unk unk unk unk unk unk unk unk unk]

Do you think the output is like that because I tested the model's performance too soon or that the model is not learning anything?

LeenaShekhar avatar Aug 07 '17 20:08 LeenaShekhar

I am facing with the exact same problem, can someone guide us.

karanpande avatar Oct 09 '17 22:10 karanpande

I have not worked on it after that, but from experience it might be because of the training. Usually you need t train for longer steps in such models. Try that and see it that works. I have not explored the pre-trained models tough. You can compare your output with that.

LeenaShekhar avatar Nov 16 '17 01:11 LeenaShekhar