attention-analysis
attention-analysis copied to clipboard
Dear Kevin, hello! I'm trying to reproduce your work described at https://github.com/clarkkev/attention-analysis but it's not clear how to get the train.txt and dev.txt files used in preprocess_depparse.py. Can you clarify...
Can you please clarify what does each number (8,10) stand for in "Head 8-10" in Figure 5 of your article, i.e. layer or head number? Thank you very much for...
1. Using unique `_` separator in preprocess_unlabeled.py (sync with README.md); 2. Providing `--num_docs` arg at README.md; 3. Fixed bug with json writing when path contains only filename.
I am trying this with tf 2.0 and am having a lot of issues. can you tell us what version you were using or maybe add a requirements.txt? .. thanks!
The instructions say: We include two pre-processing scripts for going from a raw data file to JSON that can be supplied to attention_extractor.py I think what is meant is the...
It should have been inserted at 0th index in -1 dimension, currently its added at last index. Since attention approximated for ROOT by adding start/end tokens would be at 0th...
Any plans of releasing the code for coreference analysis in the paper? Alternatively, it is possible to explain the methodology? Mainly around how the "head" word is chosen, and what...