nntrainer icon indicating copy to clipboard operation
nntrainer copied to clipboard

Support Convolution&Batchnorm Fusing for Optimized Inference Mode

Open DonghakPark opened this issue 8 months ago • 3 comments

One of the ways to accelerate running the NNTrainer in inference mode is to fuse operations. We are currently using this fusion when exporting to TensorFlow Lite. By applying it to the current NNTrainer, we can improve the speed during inference.

Many Deeplearning Model using Batchnorm after Conv layer

img1 daumcdn

and when inference we can fusing ops to below

img1 daumcdn img1 daumcdn

image ref

DonghakPark avatar Jun 19 '24 01:06 DonghakPark