scibert icon indicating copy to clipboard operation
scibert copied to clipboard

A BERT model for scientific text.

Results 62 scibert issues
Sort by recently updated
recently updated
newest added

I found `gradient_accumulation_batch_size` exists in several scibert conf, such as https://github.com/allenai/scibert/blob/8562a120e6788dcbadbe05ef7fd4463dee17ee59/allennlp_config/ner.json but allennlp trainer doesn't have this param, https://github.com/allenai/allennlp/blob/master/allennlp/training/trainer.py

I tried with the code below: ``` from transformers import AutoTokenizer, AutoModel,AutoModelForTokenClassification import torch #I am getting the label list from labels.txt file present in the Pytorch Huggingface model(scibert-scivocab-uncased) def...

I have my own data set consisting a a few hundred abstracts and I want to see baseline performance using Sci-BERT's PICO functionality. Are there code snip bits for easily...

Hi everybody, We are a group of uni students currently working on a research project within a NLP class. We extracted text from a GIEC report and build a Knowledge...

Old API `model = AutoModel.from_pretrained('allenai/scibert_scivocab_uncased')` raises exception therefore updating it to `model = AutoModelWithLMHead.from_pretrained('allenai/scibert_scivocab_uncased')`

I noticed that in the datasets of relation classification, there is no entity pair of 'N/A' label (no relation between two entities). In the other words, entities without any relations...

Hi, I am interested in the parsing field of nlp and am studying sciBERT. In parsing using genia, ValueError was raised as shown in the attached file. How can I...

I have two questions firstly:I'm running train_allennlp_local.sh with a bug: ModuleNotFoundError: No module named 'scibert' YML@Spuer-HR:~/Jiaxin/scibert-master/scripts$ ./train_allennlp_local.sh output/ 2020-11-10 11:08:03,865 - INFO - pytorch_pretrained_bert.modeling - Better speed can be achieved...

Hi, I want to pretrain SciBERT using additional data, and I want to enlarge the vocabulary with 100 additional "domain-specific" terms which are reserved for such usage. So I've figured...