tf_chatbot_seq2seq_antilm icon indicating copy to clipboard operation
tf_chatbot_seq2seq_antilm copied to clipboard

Maintaining state between predictions

Open gidim opened this issue 7 years ago • 2 comments

Hi, Any plans on adding state to the encoder/decoder? the idea is that you realistically you want to predict (answer_n | question_n, answer_n-1,question_n-1 ...) and not one by one as the original translation model is doing.

gidim avatar Mar 16 '17 17:03 gidim

That's an interesting idea, how do we make this model remembering some facts from previous dialogue? I guess neural Turing machine might be a good candidate.

Marsan-Ma-zz avatar Mar 21 '17 07:03 Marsan-Ma-zz

There's many ways to maintain some memory of the sequence of inputs but the easiest is just to keep the LSTM/GRU state between calls to model.step(), and not reset it.

gidim avatar Mar 25 '17 22:03 gidim