distiller icon indicating copy to clipboard operation
distiller copied to clipboard

How to convert from a quantization-aware training model to a post-training quantization model?

Open zhengge opened this issue 6 years ago • 5 comments

zhengge avatar Jun 17 '19 06:06 zhengge

want to ask the same question

dongzhen123 avatar Jun 17 '19 07:06 dongzhen123

It seems that converting from a quantization-aware training model to a post-training quantization model is not yet supported by the document. https://nervanasystems.github.io/distiller/algo_quantization.html

Is there any plan to do this?

robotcator avatar Aug 26 '19 03:08 robotcator

I (and I think from the Issues entries, many others as well) would also be interested :)

asti205 avatar Jan 09 '20 09:01 asti205

Hi,

Sorry for the really late response... The way QAT was implemented - the model is re-quantized each minibatch run. So, in the end of the training, the model is actually already quantized and ready to use, i.e. the .weights are already quantized and so are the activations. No need to PostTrainQuantize the model.

levzlotnik avatar Feb 19 '20 12:02 levzlotnik

@zhengge @dongzhen123 could you please elaborate what you mean when you say "convert the model", and also what do you aim to accomplish from this?

Thanks.

shazib-summar avatar Jun 18 '20 08:06 shazib-summar