Iz Beltagy
Iz Beltagy
You might also want to give Longformer a shot, especially if you are working on an NLP task as it includes a pretrained model for long docs https://github.com/allenai/longformer (self-promotion :D)
It is just one function call: https://github.com/allenai/scibert/blob/master/scripts/cheatsheet.txt#L6 The output format is slightly different than what BERT expects, so we manually fixed after it was generated.
I am not familiar with how HF TF support works, but as far as I understand, we don't need to do anything specific on the model side to make it...
Yes, AllenNLP doesn't support gradient accumulation. We have it implemented in our fork of allennlp (check requirements: https://github.com/allenai/scibert/blob/master/requirements.txt)
@InesArous, you can try to follow one of the classification examples in the HF code https://github.com/huggingface/transformers/tree/master/examples/text-classification, maybe the `run_pl_glue.py` one.
@amandalmia14, you need to use `AutoModelForSequenceClassification ` instead of `AutoModel`
It is just English.
@shreyashub, I think you are talking about bc5cdr not JNLPBA because JNLPBA doesn't have Disease category. For bc5cdr, we used a version that we had in s2 that dropped the...
sorry for the confusion. citation_intent: SciCite mag: Paper Field
This is an AllenNLP issue. Can you share the error stack trace?