Pretrained-Language-Model icon indicating copy to clipboard operation
Pretrained-Language-Model copied to clipboard

TinyBert fine tune SQuAD

Open shairoz-deci opened this issue 4 years ago • 2 comments

Thank you for sharing this great repo. Can you please provide instructions, or code if available, for task distillation on the SQuAD dataset?

Thanks in advance

shairoz-deci avatar Aug 16 '21 12:08 shairoz-deci

Hi, we have no plans to release the distillation code for SQuAD dataset. For SQuAD fine-tune code, you can refer to https://github.com/huawei-noah/Pretrained-Language-Model/blob/master/AutoTinyBERT/superbert_run_en_classifier.py.

zwjyyc avatar Sep 17 '21 02:09 zwjyyc

thank you for your reply @zwjyyc I've implemented the squad training based on the example you mentioned, can perhaps verify that the training of squad was done similarly to the glue tasks? meaning that the data was multiplied by 20 using augmentation and ran for 10 epochs without pred_distill and 3 more epochs with pred_distil? By doing the above the training time seems very long, almost as long as the general distillation.

Thanks

shairoz-deci avatar Sep 17 '21 10:09 shairoz-deci