distiller
distiller copied to clipboard
How to convert from a quantization-aware training model to a post-training quantization model?
want to ask the same question
It seems that converting from a quantization-aware training model to a post-training quantization model is not yet supported by the document. https://nervanasystems.github.io/distiller/algo_quantization.html
Is there any plan to do this?
I (and I think from the Issues entries, many others as well) would also be interested :)
Hi,
Sorry for the really late response...
The way QAT was implemented - the model is re-quantized each minibatch run. So, in the end of the training, the model is actually already quantized and ready to use, i.e. the .weights are already quantized and so are the activations. No need to PostTrainQuantize the model.
@zhengge @dongzhen123 could you please elaborate what you mean when you say "convert the model", and also what do you aim to accomplish from this?
Thanks.