Results 4 issues of Kazutoshi Shinoda

https://github.com/jojonki/BiDAF/blob/3e5ac9c76d02de2d8f75b1eda6632f8a9432eba6/layers/char_embedding.py#L28 I feel strange to see this code. Why do you sum over `word_len` dimension? Why don't you apply 1D filter over `word_len` dimension? Thank you.

https://github.com/lena-voita/description-length-probing/blob/2696af04226cff191eff265dbed0744512bad7b5/control_tasks/control_tasks/data.py#L325 Regarding this function, I found the following error case. Even though this may be a minor error, just for your information. ``` from transformers import AutoTokenizer # preparing an...

https://github.com/ZiJianZhao/SeqGAN-PyTorch/blob/master/main.py#L87 why? In the original implementation by LantaoYu, exp is not used for loss values.