scibert
scibert copied to clipboard
A BERT model for scientific text.
where is the result on dataset citation_intent and mag in paper?
Hi, I am reproducing the fine-tune results following your instruction. I work on your default code and my setting is as below. I work on ebmnlp and pico task without...
Hi, I noticed something weird about the `max_len` attribute of the `tokenizer` ```python from transformers import AutoTokenizer tokenizer = AutoTokenizer.from_pretrained("allenai/scibert_scivocab_uncased") print(tokenizer.max_len) # => 1000000000000000019884624838656 ``` Whereas I expected it to...
https://github.com/allenai/scibert/blob/ae9f98f58a9793781148ccf715b03f7160060f24/scibert/models/bert_text_classifier.py#L37 BertTextClassifier requires an instance of TextFieldEmbedder and calls `get_output_dim` on that instance, however that method always raises NotImplementedError. Is this meant to deliberately un-implement a method from the superclass,...
https://github.com/allenai/scibert/blob/ae9f98f58a9793781148ccf715b03f7160060f24/scibert/models/bert_text_classifier.py#L27 I believe that this line should read `verbose_metrics: bool = False`. As it stands, it doesn't have a default value.
Traceback (most recent call last): File "/home/hui2019/scibert-master/scibert-master/src/allennlp/allennlp/common/util.py", line 289, in get_spacy_model spacy_model = spacy.load(spacy_model_name, disable=disable) File "/home/hui2019/anaconda3/envs/py36_pytorch1.3.1/lib/python3.6/site-packages/spacy/__init__.py", line 27, in load return util.load_model(name, **overrides) File "/home/hui2019/anaconda3/envs/py36_pytorch1.3.1/lib/python3.6/site-packages/spacy/util.py", line 139, in load_model...
`` import tensorflow as tf from transformers import * #BertTokenizer, TFBertModel tokenizer = AutoTokenizer.from_pretrained('allenai/scibert_scivocab_uncased') model = AutoModel.from_pretrained('allenai/scibert_scivocab_uncased') input_ids = tf.constant(tokenizer.encode("My dog is cute"))[None, :] # Batch size 1 outputs =...
Hi, do you plan to make the pretraining corpus available, or provide a way to reproduce / approximate it using Semantic Scholar?
We'd like to build a specific RE model to meet our own requirement. The performance is quite good. However, we identified some overfitting problems and we'd like to solve the...
 what's the wrong,I only modify the path .