model-compression topic
OFQ
The official implementation of the ICML 2023 paper OFQ-ViT
Lossless_Compression
We propose a lossless compression algorithm based on the NTK matrix for DNN. The compressed network yields asymptotically the same NTK as the original (dense and unquantized) network, with its weight...
knowledgeDistillation
PyTorch implementation of (Hinton) Knowledge Distillation and a base class for simple implementation of other distillation methods.
Compressed-Transformers
In this repository, we explore model compression for transformer architectures via quantization. We specifically explore quantization aware training of the linear layers and demonstrate the performanc...
zsdb3kd
Knowledge distillation (KD) from a decision-based black-box (DB3) teacher without training data.
lm-vocab-trimmer
Vocabulary Trimming (VT) is a model compression technique, which reduces a multilingual LM vocabulary to a target language by deleting irrelevant tokens from its vocabulary. This repository contains a...
Xcompression
[ICLR 2022] Code for paper "Exploring Extreme Parameter Compression for Pre-trained Language Models"(https://arxiv.org/abs/2205.10036)
BiFSMNv2
Pytorch implementation of BiFSMNv2, TNNLS 2023
octopus-ml
A collection of handy ML and data visualization and validation tools. Go ahead and train, evaluate and validate your ML models and data with minimal effort.
SigDL
Deep Learning Compression and Acceleration SDK -- deep model compression for Edge and IoT embedded systems, and deep model acceleration for clouds and private servers