Knowledge-Distillation-in-Keras
Knowledge-Distillation-in-Keras copied to clipboard
Demonstrates knowledge distillation for image-based models in Keras.
Knowledge-Distillation-in-Keras
Demonstrates knowledge distillation (kd) for image-based models in Keras. To know more check out my blog post Distilling Knowledge in Neural Networks that accompanies this repository. The blog post covers the following points -
- What is softmax telling us?
- Using the softmax information for teaching - Knowledge distillation
- Loss functions in knowledge distillation
- A few training recipes
- Experimental results
- Conclusion
About the notebooks
Distillation_Toy_Example.ipynb- kd on the MNIST datasetDistillation_with_Transfer_Learning.ipynb- kd (with the typical KD loss) on the Flowers dataset with a fine-tuned modelDistillation_with_Transfer_Learning_MSE.ipynb- kd (with an MSE loss) on the Flowers dataset with a fine-tuned modelEffect_of_Data_Augmentation.ipynb- studies the effect of data augmentation on kd
Results
Interact with the all the results here.
Acknowledgements
I am grateful to Aakash Kumar Nain for providing valuable feedback on the code.
