MyDecember12

Results 3 comments of MyDecember12

Thank you for your reply. First of all, I used the latest version of amirstan_plugin, and then the fp16 mode I used was effective, because I compared the size before...

And I want to ask how much the size of your converted model is reduced, and how much the inference time is reduced

Thank you very much, because these problems have troubled me for a long time. My device is Tesla V100. The convert and inference codes are as follows: ``` import numpy...