compress-net-notes icon indicating copy to clipboard operation
compress-net-notes copied to clipboard

This is a collection of papers aiming at reducing model sizes or the ASIC/FPGA accelerator for Machine Learning, especially deep neural network related applications. (Inspired by Embedded-Neural-Network.)

You can use the following materials as your entrypoint:

Terminologies

  • Structural pruning (compression): compress CNNs based on removing "less important" filter.

Network Compression

Reduce Precision

Deep neural networks are robust to weight binarization and other non-linear distortions showed that DNN can be robust to more than just weight binarization.

Linear Quantization

Non-linear Quantization

Reduce Number of Operations and Model Size

Exploiting Activation Statistics

  • To be updated.

Network Pruning

Network Prune: a large amount of the weights in a network are redundant and can be removed (i.e., set to zero).

Bayesian network pruning

  • [1711]. Interpreting Convolutional Neural Networks Through Compression - [notes][arXiv]
  • [1705]. Structural compression of convolutional neural networks based on greedy filter pruning - [notes][arXiv]

Compact Network Architectures

Knowledge Distillation

A Bit Hardware

Contributors