tensorrt icon indicating copy to clipboard operation
tensorrt copied to clipboard

TensorRT saved model too large to use with TFServing

Open bharatv007 opened this issue 4 years ago • 4 comments

Versions: Tensorflow- 2.3.0-rc1 CUDA-10 TensorRT-6 I am trying to convert a GPT2 model, the saved model size is about 1.9GB. It causes an issue when I try to use TF serving for deployment as it hits a protobuf limit of 1 GB. I have tried to not build TRT engines before deployement too, but it did not affect the size of the saved_model.pb.

bharatv007 avatar Jul 22 '20 20:07 bharatv007

CC @bixia1

sanjoy avatar Jul 24 '20 01:07 sanjoy

Any updates on this? @sanjoy @bixia1

1vn avatar Jul 30 '20 15:07 1vn

I met the same problem on tf-2.4.1

ericxsun avatar Mar 03 '21 03:03 ericxsun

I also met this problem

forrest0402 avatar Jul 29 '21 11:07 forrest0402