model-optimization icon indicating copy to clipboard operation
model-optimization copied to clipboard

A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.

Results 228 model-optimization issues
Sort by recently updated
recently updated
newest added

Hi team `tensorflow-model-optimization`, I was applying clustering and pruning to a model based on Bert encoder from `tensorlfow-models-official`, and noticed that there is no API for registering custom layers, so...

technique:pruning
technique:clustering
api-review

Does Post-training full integer quantization in https://www.tensorflow.org/lite/performance/post_training_integer_quant#convert_using_float_fallback_quantization support BERT? I convert my pb model to tf lite: ``` dataset = create_dataset() def representative_dataset(): for data in dataset: yield { "token_type_ids":...

bug

I've read through the official guide and ran into problems understanding some concepts: 1. Is it possible to use Quantization Aware Training and not convert the model to a TF...

**Describe the bug** Stripping the pruning layers seems to somehow disconnect the input layer from the graph. **System information** TensorFlow version (installed from source or binary): 2.11 (macos) TensorFlow Model...

bug

Prior to filing: check that this should be a bug instead of a feature request. Everything supported, including the compatible versions of TensorFlow, is listed in the overview page of...

bug

use quantize_model interface: ![image](https://user-images.githubusercontent.com/19814680/232799979-c3c83382-78bb-4dcf-bf9b-9abb6216cc4e.png) original convert interface: ![image](https://user-images.githubusercontent.com/19814680/232800108-0a06b90c-1d80-47e9-a29d-0bbc2f95b002.png)

bug

Hi all. I've recently trained a keras implementation of ssd-keras. I've managed to run QAT training on the model and got desired the accuracy. I wanted to get the quantised...

**Describe the bug** When using `tf.keras.mixed_precision.experimental.Policy("mixed_float16", loss_scale="dynamic")` the `sparsity.prune_low_magnitude` fails in tensor type conversion with the error `Tensor conversion requested dtype float32 for Tensor with dtype float16: `. Things work...

feature request
technique:pruning

Remove extra parentheses in line 726 to fix the broken link.

technique:pruning