deepnlp icon indicating copy to clipboard operation
deepnlp copied to clipboard

[textsum]Attention heatmap error

Open JuruoMP opened this issue 7 years ago • 2 comments

Compared with the headmap showing in "A Neural Attention Model for Abstractive Sentence Summarization"(A.M.Rush), there are some problems here.

A heatmap should map the attention relationship between input and output, which means that it should have a higher attention if the output word is copied from input sentence. However, the heatmap on this project is just a mass, which is hard to get a point of which input word attentions which output word.

JuruoMP avatar Jun 02 '17 06:06 JuruoMP

Hi, Thanks for your feedback. The method in the project is a little bit tricky and just a short-term solution to get the Attn-Mask tensor out (tensorflow 1.0). Since tensorflow 1.2 already deleted the old archived RNN functions and changed to a more common dynamic_rnn() function. I will find a more elegent way to do that in the new tf version. Any ideas are very welcome.

rockingdingo avatar Jul 19 '17 11:07 rockingdingo

@rockingdingo Thanks for replying. And I have another question. Lots of methods of visualization has been focused recently, such as sensitivity analysis and relevance propogation. In which method are you going to use to generate the heat map?

JuruoMP avatar Jul 25 '17 14:07 JuruoMP