aimet
aimet copied to clipboard
quantsim.export(path, filename_prefix) could not generate int8 QNN ONNX model
After calling quantsim.export(path, filename_prefix), I could not get int8 QNN ONNX model. My objective is to get an int8 ONNX model through aimet quant toolkit, which shows like the attached image below.
However, by calling quantsim.export(path, filename_prefix), I only can get pth files, encoding files and one fp32 ONNX model. Did I use the export functionality incorrectly? Or is any way to convert encoding files and the fp32 ONNX model to one int8 QNN model?
You used it correctly, you can take the encodings and FP32 model to a quantized target to get a quantized model. AIMET only simulates HW performance
@JiliangNi please use --keep_quant_nodes option with the qnn converters to see a QNN model with activation quant/dequant nodes. Without this option, quant nodes are stripped from the graph.