neural-compressor topic
List
neural-compressor repositories
optimum-benchmark
201
Stars
33
Forks
Watchers
A unified multi-backend utility for benchmarking Transformers, Timm, PEFT, Diffusers and Sentence-Transformers with full support of Optimum's hardware optimizations & quantization schemes.
auto-round
81
Stars
9
Forks
Watchers
SOTA Weight-only Quantization Algorithm for LLMs. This is official implementation of "Optimize Weight Rounding via Signed Gradient Descent for the Quantization of LLMs"