BERT-pytorch icon indicating copy to clipboard operation
BERT-pytorch copied to clipboard

Google AI 2018 BERT pytorch implementation

Results 66 BERT-pytorch issues
Sort by recently updated
recently updated
newest added

The position embedding in the BERT is not the same as in the transformer. Why not use the form in bert?

good first issue

if I want to do NER,so What should I do with your code

The base BERT model in https://arxiv.org/pdf/1810.04805.pdf uses 768 hidden features, 12 layers, 12 heads (which are also the defaults in `bert.py`), while the default configuration in the argparser of `__main__.py`...

I'm relatively new to pytorch and BERT, but I figured out how to pretrain the BERT model on my dataset. I want to use the model to predict the mask(s)...

I try to run the code on a small dataset and I find that pred_loss decrease fast while avg_acc stay at 50%. It is strange to me since decrease in...

help wanted
invalid
question

Hi! I am new to pytorch. So, could you provide a small example to illustrate how to use this pre-trained model to load, inference and finetune? Thx

Predicting 'is_next' seems incorrect. Since the 'is_next' label value consists of 0 and 1, ignore_index should be 'None'.

I want to output the word vector