Yufan He
Yufan He
model inference server mean trtis (triton) ? TRT plan has has some constraints: > Note: The generated plan files are not portable across platforms or TensorRT versions. Plans are specific...
There are roughly two ways to consider: - Transform all models into a unified type (eg onnx. Use onnxruntime to provide inference). At this time, the task needs some resources...
> > Transform all models into a unified type (eg onnx. Use onnxruntime to provide inference). At this time, the task needs some resources and dependencies. It is recommended to...
@nautilusshell Please provide more information then reopen this issue again.
Thx for your reply. I think if Neuropod only support python SDK (with out use_native), supporting newly frameworks maybe easier. By the way where can I find Neuropod 's roadmap...