rcnn-relation-extraction
rcnn-relation-extraction copied to clipboard
Macro-F1 score is lower than anticipated
http://nlpprogress.com/relationship_extraction.html mentions this GitHub as a reimplementation of the following paper: http://www.aclweb.org/anthology/P16-1072 which achieves a macro F1 of 86.3. Upon running the model off the shelf, we achieve a macro F1 of 0.51. Running the eval script shows a macro F1 of 0.48. Is there something that I am missing in running this code? I am running default hyperparameter settings; is there an ideal set of hyperparameter settings?
@geetickachauhan Hi, sorry for late. I think there are some mistakes. This repository is not reimplementation of that paper. This is a project that applied Lai's rcnn(https://www.aaai.org/ocs/index.php/AAAI/AAAI15/paper/view/9745) to the relation extraction task.
@sebastianruder Hi, Ruder. Nice to meet you! I found out that my project was mentioned in your post through @geetickachauhan. It is a great honor for introducing my project through your wonderful post :+1: : http://nlpprogress.com/english/relationship_extraction.html. However, my project is not reimplementation of BRCNN (Cai et al., 2016), unfortunately. I hope you can fix it and improve the completeness of your post. PS. there are some reimplementation projects for relation extraction: CNN (https://github.com/roomylee/cnn-relation-extraction) and Attention-BiLSTM (https://github.com/SeoSangwoo/Attention-Based-BiLSTM-relation-extraction). I would really appreciate it if you introduce them to your post together. :smile:
Hi Joohong, thanks a lot for the note! Could you add a PR to correct the entry in the table? I think you're better able to describe your repo than me. 😄Also feel free to add the other reimplementation links, as I'm not sure which models they are associated with.
I will make a pull request, ASAP! Thanks a lot, Ruder. :smile:
I add a PR. Please check it. Thanks! @sebastianruder 👍
Thanks for updating this, everyone! Also @roomylee have you noticed an issue with re-implementing papers that claimed to be achieving macro F1's of 86-88%? The performance for one such re-implementation by my lab mate was in the low 80s.
@geetickachauhan If hyper-parameter tuning is done well, it may be overcome. L2 regularization may help you, too. It is good to refer to my cnn-relation-extraction and SeoSangwoo's Attention-Based-BiLSTM-relation-extraction.
Thanks for the PR (https://github.com/sebastianruder/NLP-progress/pull/144), @roomylee! :)