distiller
distiller copied to clipboard
Quantization Capabilities in PyTorch
Hi Team,
Thanks for the great tool for model compression research.
As it is stated that you guys are planning to add PTQ or capability to export the quantized model to ONNX, would be greatly interested in knowing if that's yet in the pipeline, or it's better to consider PyTorch internal quantization capabilities or using TensorRT for the same.
Thanks!
A formal method or pipeline to export models after PTQ would be awesome