Anqi Liu
Anqi Liu
@ihaohe Could you please provide the version information for your CUDA and TensorRT?
Hi I checked your log. This appears to be a model with a quantize-dequantize (qdq) node. `Error[4]: DLA Standalone is enabled but layer: '/model.0/conv/**_input_quantize**r/Constant_1_output_0' is not supported on DLA. `...
Hi @tangxiangyu0313 , The code in this [link](https://github.com/NVIDIA-AI-IOT/cuDLA-samples/tree/main/export/yolov5-qat) has already implemented quantization for YOLOv5. We provide two quantization modes for selection, and you can refer to the[ README](https://github.com/NVIDIA-AI-IOT/cuDLA-samples/blob/main/export/README.md) for replication....
@zyitom It seems that the library installation was not successful. You can try using the following command to install it. ```shell $ git clone -b v8.6.1 https://github.com/NVIDIA/TensorRT.git $ cd TensorRT/tools/pytorch-quantization...
If we use https://github.com/NVIDIA-AI-IOT/cuDLA-samples/tree/main/export#option1, the generated model can also run on the GPU. However, If the Q&DQ nodes of these tensors are inconsistent, there are a lot of useless int8->fp16...