ALBERT-Pytorch icon indicating copy to clipboard operation
ALBERT-Pytorch copied to clipboard

Pytorch Implementation of ALBERT(A Lite BERT for Self-supervised Learning of Language Representations)

Results 7 ALBERT-Pytorch issues
Sort by recently updated
recently updated
newest added
trafficstars

if i wanna use my own textual data to pre-train a electra from scatch, what is the format of the text? Only sentence segmentation or even more ?? Please help.

hey there, sadly i dont clearly understand how the wiki.train.tokens and vocab.txt file is produced. followingly could you please show me the steps on how to reproduce the same train.tokens...

Is there a bug in there?

what should be my mask_alpha and mask_beta values if my seqence lenght is about 10-20?

i am trying to train albert model using my own dataset i don't know why got the error like this please let me know, when you figure out thank you

I'm running classify on the MRPC dataset. In trainer.train trainer.train(get_loss,model_file,True), it allows only three arguments not 4 so I cant use the pretrain file. Also it runs out of memory,...

In ALBERT(Lan at el), There is not detail about 80% mask ![image](https://user-images.githubusercontent.com/10525011/66101701-ba2b0400-e5ea-11e9-9375-5aeca1a3173e.png) But, from n-gram masking (Joshi et al., 2019), they said about 80/10/10 > As in BERT, we also...