model-compression topic
SVD-LLM
Official Code for "SVD-LLM: Truncation-aware Singular Value Decomposition for Large Language Model Compression"
logit-standardization-KD
[CVPR 2024 Highlight] Logit Standardization in Knowledge Distillation
LumiNet
The official implementation of LumiNet: The Bright Side of Perceptual Knowledge Distillation https://arxiv.org/abs/2310.03669
Pruning-LLMs
The framework to prune LLMs to any size and any config.
OpenBA-v2
OpenBA-V2: 3B LLM (Large Language Model) with T5 architecture, utilizing model pruning technique and continuing pretraining from OpenBA-15B.
Awesome-Diffusion-Distillation
A list of papers, docs, codes about diffusion distillation.This repo collects various distillation methods for the Diffusion model. Welcome to PR the works (papers, repositories) missed by the repo.
picollm
On-device LLM Inference Powered by X-Bit Quantization
MoA
The official implementation of the paper <MoA: Mixture of Sparse Attention for Automatic Large Language Model Compression>
Awesome-Tensor-Decomposition
😎 A curated list of tensor decomposition resources for model compression.
Awesome-Token-level-Model-Compression
📚 Collection of token-level model compression resources.