DeepSpeedExamples icon indicating copy to clipboard operation
DeepSpeedExamples copied to clipboard

How to use Deepspeed checkpoints for BingBertSquad finetuning?

Open xycforgithub opened this issue 2 years ago • 2 comments

Hi there, The tutorial https://www.deepspeed.ai/tutorials/bert-finetuning/#loading-huggingface-and-tensorflow-pretrained-models makes clear how to load HF and TF checkpoints into Deepspeed. What if we want to load a Deepspeed checkpoint, like from the Bing BERT example? Is it that we load the "mp_rank_00_model_states.pt" file in the checkpoint?

I'm currently using fp16 and ZERO-2, so I wonder if using that will lose some precision. Should I use zero_to_fp32 to convert the checkpoint to fp32 for loading?

xycforgithub avatar Aug 19 '21 22:08 xycforgithub