model-optimization icon indicating copy to clipboard operation
model-optimization copied to clipboard

some questions about quantization in TensorFlow

Open rthenamvar opened this issue 2 years ago • 0 comments

I've read through the official guide and ran into problems understanding some concepts:

  1. Is it possible to use Quantization Aware Training and not convert the model to a TF Lite model at the end?
  2. Can I change the framework's default of 8-bit quantization? In the official document 4-bit and 16-bit quantizations were mentioned as experimental meaning the models cannot be converted to TF lite models. But isn't it possible to use the models without converting them to TF Lite models?

Thanks

rthenamvar avatar Apr 26 '23 09:04 rthenamvar