TensorRT
TensorRT copied to clipboard
why plugin?
I recently tried to write a plugin, and it was a pain to write it. Even after I wrote it, there were still a lot of problems when compiling it, and it might be different on different machines. Why can't I skip the arithmetic that can't turn trt and just use the original or use the cuda kernel written by the user himself? The threshold for using plugin is indeed a bit high.
I'm developing a tool that can produce the TRT plugin from a simple config file, so the user only needs to focus on the kernel implementation of the operator :)
closing since no activity for more than 3 weeks, please reopen if you still have question, thanks!
the tool is public now: https://github.com/NVIDIA-AI-IOT/tensorrt_plugin_generator
the tool is public now: https://github.com/NVIDIA-AI-IOT/tensorrt_plugin_generator
Thanks, I'll give it a try