onnxruntime_backend icon indicating copy to clipboard operation
onnxruntime_backend copied to clipboard

Will onxxruntime backend support INT8 on cpu ?

Open bharadwajymg opened this issue 1 year ago • 1 comments
trafficstars

Hi, we are trying to quantise our onnx models to int8 to run on cpu using : https://onnxruntime.ai/docs/performance/model-optimizations/quantization.html#quantization-on-gpu

we are using dynamic quantisation , and banking on AVX2 and AVX512 extensions , when we tested our models using onnx runtime we see an improvement so cross checking if this backend supports them by directly defining backend in config.pbtxt ?

bharadwajymg avatar Feb 20 '24 15:02 bharadwajymg

yes, onxxruntime backend support INT8 on cpu

Jackiexiao avatar Mar 08 '24 07:03 Jackiexiao