scibert icon indicating copy to clipboard operation
scibert copied to clipboard

A BERT model for scientific text.

Results 62 scibert issues
Sort by recently updated
recently updated
newest added

where is the result on dataset citation_intent and mag in paper?

Hi, I am reproducing the fine-tune results following your instruction. I work on your default code and my setting is as below. I work on ebmnlp and pico task without...

Hi, I noticed something weird about the `max_len` attribute of the `tokenizer` ```python from transformers import AutoTokenizer tokenizer = AutoTokenizer.from_pretrained("allenai/scibert_scivocab_uncased") print(tokenizer.max_len) # => 1000000000000000019884624838656 ``` Whereas I expected it to...

https://github.com/allenai/scibert/blob/ae9f98f58a9793781148ccf715b03f7160060f24/scibert/models/bert_text_classifier.py#L37 BertTextClassifier requires an instance of TextFieldEmbedder and calls `get_output_dim` on that instance, however that method always raises NotImplementedError. Is this meant to deliberately un-implement a method from the superclass,...

https://github.com/allenai/scibert/blob/ae9f98f58a9793781148ccf715b03f7160060f24/scibert/models/bert_text_classifier.py#L27 I believe that this line should read `verbose_metrics: bool = False`. As it stands, it doesn't have a default value.

Traceback (most recent call last): File "/home/hui2019/scibert-master/scibert-master/src/allennlp/allennlp/common/util.py", line 289, in get_spacy_model spacy_model = spacy.load(spacy_model_name, disable=disable) File "/home/hui2019/anaconda3/envs/py36_pytorch1.3.1/lib/python3.6/site-packages/spacy/__init__.py", line 27, in load return util.load_model(name, **overrides) File "/home/hui2019/anaconda3/envs/py36_pytorch1.3.1/lib/python3.6/site-packages/spacy/util.py", line 139, in load_model...

`` import tensorflow as tf from transformers import * #BertTokenizer, TFBertModel tokenizer = AutoTokenizer.from_pretrained('allenai/scibert_scivocab_uncased') model = AutoModel.from_pretrained('allenai/scibert_scivocab_uncased') input_ids = tf.constant(tokenizer.encode("My dog is cute"))[None, :] # Batch size 1 outputs =...

Hi, do you plan to make the pretraining corpus available, or provide a way to reproduce / approximate it using Semantic Scholar?

We'd like to build a specific RE model to meet our own requirement. The performance is quite good. However, we identified some overfitting problems and we'd like to solve the...

![image](https://user-images.githubusercontent.com/49578851/77167053-3f9a6600-6af0-11ea-8907-2744e1142520.png) what's the wrong,I only modify the path .