electra icon indicating copy to clipboard operation
electra copied to clipboard

Using own data to continue pre-training from the released ELECTRA checkpoints

Open ghost opened this issue 4 years ago • 4 comments

Hello,

I would like to pretrain using my own corpus but starting from the released Electra model weights. In the Readme file, it says I should download the pretrained weights into $DATA_DIR/electra_small if I am to use the small model.

Is this a typo?

I think I should download the weights into $DATA_DIR/models/electra_small instead, so that it will look into models/ folder to continue training when I run the run_pretraining.py.

If I follow the Readme file, downloading weights into $DATA_DIR/electra_small, I am actually pretraining the small model from scratch using my own data, with small model structure. Is my understanding correct?

Thank you in advance for the advice.

ghost avatar Sep 17 '20 19:09 ghost

same question here, I try to pretrain from original electra small model weights, but i get ERROR:tensorflow:Error recorded from training_loop: Restoring from checkpoint failed. This is most likely due to a Variable name or other graph key

hyusterr avatar Dec 10 '20 18:12 hyusterr

I have the same problem

xieexiaotuzi avatar Mar 24 '22 12:03 xieexiaotuzi

I am also struggling to pretrain starting from a pre-built model. The instructions in the README are not working.

Joseph-Vineland avatar May 02 '22 06:05 Joseph-Vineland

I am also struggling to pretrain starting from a pre-built model. The instructions in the README are not working.

Agree. I think the instructions in the README only indicate that you can train more steps from scratch. Any ideas on continuing training from the previous checkpoints?
Thank you for any hints and ideas.

JiazhaoLi avatar Sep 22 '22 01:09 JiazhaoLi